Skip to main content

ORIGINAL RESEARCH article

Front. Plant Sci., 03 February 2023
Sec. Plant Biophysics and Modeling
This article is part of the Research Topic Plant Architectural Models and Crop Production View all 8 articles

Deep learning for crown profile modelling of Pinus yunnanensis secondary forests in Southwest China

  • 1School of Environmental and Resources Science, Zhejiang A & F University, Hangzhou, China
  • 2Department of Mathematics and Computer Science, Dali University, Dali, Yunnan, China

Accurate information concerning crown profile is critical in analyzing biological processes and providing a more accurate estimate of carbon balance, which is conducive to sustainable forest management and planning. The similarities between the types of data addressed with LSTM algorithms and crown profile data make a compelling argument for the integration of deep learning into the crown profile modeling. Thus, the aim was to study the application of deep learning method LSTM and its variant algorithms in the crown profile modeling, using the crown profile database from Pinus yunnanensis secondary forests in Yunnan province, in southwest China. Furthermore, the SHAP (SHapley Additive exPlanations) was used to interpret the predictions of ensemble or deep learning models. The results showed that LSTM’s variant algorithms was competitive with traditional Vanila LSTM, but substantially outperformed ensemble learning model LightGBM. Specifically, the proposed Hybrid LSTM-LightGBM and Integrated LSTM-LightGBM have achieved a best forecasting performance on training set and testing set respectively. Furthermore, the feature importance analysis of LightGBM and Vanila LSTM presented that there were more factors that contribute significantly to Vanila LSTM model compared to LightGBM model. This phenomenon can explain why deep learning outperforms ensemble learning when there are more interrelated features.

1 Introduction

Pinus yunnanensis is a major component of coniferous forests in southwestern China. It has been extensively cultivated for reforestation and ecological engineering (Sun et al., 2008). In southwest China, it occupies approximately 52% of the forested area and produces 32% of the timber volume (Jin and Peng, 2004). As a pioneer tree species, Pinus yunnanensis is shade-intolerant, deep-rooted, drought-resistant, and tolerant of rocky soils with low fertility (Xu et al., 2016). It plays a crucial role in regional economic development and ecological restoration (Wu, 1986; Jin and Peng, 2004; Xu et al., 2016). Knowledge of the dynamics of Pinus yunnanensis promote the regeneration of secondary forests and further help China to better achieve its carbon neutrality target in 2060 because it can effectively improve forest carbon sinks in the region (Deng et al., 2022).

Crown profile is mostly related to the competition of individual trees in the stands, light interception, growth, and yield of trees. However, crown measurement is time-consuming and labor-intensive. It is impossible to measure the crown of every tree in actual production, so it is necessary to build the high-precision crown profile models. Crown profile models are key components of growth and yield models in the evaluation of competition among trees, forest microclimate, and biodiversity (Dong et al., 2016; Sun et al., 2022). Accurate information concerning crown profile is critical in analyzing biological processes (e.g., photosynthesis, stand growth, survival, and competition) (Rautiainen et al., 2008; Dong et al., 2016; Wang et al., 2021), whilst providing a more accurate estimate of carbon balance (Plesoianu et al., 2020; Yang et al., 2022). In mixed species ecosystems, modeling individual specie crown profile models is necessary if seeking to analyse species-specific phenological trends, plasticity, and responses to extreme events (Fawcett et al., 2021).

In forest surveys, the crown profile database contains multiple measurements for each sample tree crown, which displays hierarchical structural features. Crown profile models were initially fit assuming that the within equation errors were independent and identically distributed. However, since crown profile models in that several measures are taken on each crown profile, autocorrelation among measures within a profile is likely. For linear statistical models, the least squares estimate of regression coefficients remain unbiased and consistent in the presence of autocorrelation, but they are no longer efficient (Myers, 1990; Crecente-Campo et al., 2009). It has been of interest for forest modelers to better understand this phenomenon, particularly on the basis of statistical modeling and analysis (Wang et al., 2017). The traditional crown profile modeling methods have been used to deal with the autocorrelation and heteroscedasticity in the crown profile equations, they are mainly direct variance-covariance modelling (Hann, 1999; Crecente-Campo et al., 2009; Crecente-Campo et al., 2013), mixed-effects modelling (Fu et al., 2013; Sharma et al., 2016; Fu et al., 2017; Sharma et al., 2017; Sun et al., 2017; Jia and Chen, 2019; Wang et al., 2019; Chen et al., 2021; Di Salvatore et al., 2021), and nonlinear marginal modeling (McCulloch and Searle, 2001; Lejeune et al., 2009; de-Miguel et al., 2012; Chen et al., 2022). With the rapid development of machine learning artificial intelligence, some machine learning algorithms have the characteristics of high accuracy and good robustness for the data with nonlinear features (Singh et al., 2016; Dong et al., 2021), which has subsequently been applied to crown profile modeling. Tian et al. (2021) established crown profile model for Chinese fir (Cunninghamia lanceolata (Lamb.) Hook) based on random forest algorithm, the accuracy of the random forest model was higher than that of the mathematical model. Chen et al. (2022) recently proposed six machine learning algorithms (MLP, SVR, RF, AdaBoost, GBDT and XGBoost) for the crown profile model of China fir, and found that the performance of the ensemble learning algorithms were superior to single machine learning algorithms and parametric regression approach. However, it appeared that none of the machine learning crown profile modeling methods offered plausible explanation of hierarchical structural features. Here we are facing with space-evolving multidimensional structures (the crown profile database). Deep learning, which refers to machine learning algorithms that construct hierarchical architectures of increasing sophistication (Reichstein et al., 2019), has achieved notable success in modelling ordered sequences and data with spatial context in many fields. Applications to problems in crown profile modeling are in their infancy, but across the key problems (regression, space- or time-dependent data prediction) there are promising. LSTM network can fully explore the internal correlation between time series data, which is specially used to solve the problem of long-term information dependence and avoid gradient disappearance or explosion (Wu, 2019). At present, the application of LSTM model is only based on single factor prediction, and most of them are applied in small sample range. The similarities between the types of data addressed with classical deep learning applications and crown profile data make a compelling argument for the integration of deep learning into the crown profile modelling. There are few reports on the application of crown profile research based on deep learning.

The objectives of this study were to explore the application of deep learning method LSTM and its variant algorithms in the crown profile modeling, using the crown profile database from Pinus yunnanensis secondary forests in Yunnan province, in southwest China. It is expected to overcome many of the limitations that have hindered a more wide-spread adoption of machine learning in crown profile modeling problem.

The principal contributions of this paper are as follows: (1) A deep learning prediction based on LSTMs is introduced to explore and exploit the implicit information of hierarchical structural features for crown profile forecasting; (2) To improve the generalization capability and robustness of a single deep learning approach, LSTM’s variant algorithms consisting of a cluster of LSTMs with diverse hidden layers and neurons and LightGBM are developed; (3) A unified framework SHAP were adopted to interprete predictions of ensemble and deep learning models;(4) The performance of the proposed LSTM’s variant algorithms is successfully validated on studies data collected from Pinus yunnanensis secondary forests in Yunnan province. Statistical tests of experimental results have demonstrated the proposed LSTM’s variant algorithms is competitive with traditional Vanila LSTM, but substantially outperform ensemble learning model LightGBM.

Notation

The following notations (Table 1) will be used throughout the remainder of this paper.

TABLE 1
www.frontiersin.org

Table 1 Notation.

2 Materials and methods

2.1 Data source and processing

The localities where forest inventories were carried out at Cangshan Mountain(25°34′~26°00′N,99°55′~100°12′E), Yunnan province, SW China, including Malong Peak, Foding Peak and Maer Peak. The three peaks are located on the eastern slope of Cangshan Mountain. The eastern slope of Cangshan Mountain belongs to subtropical climate, with an annual average temperature of 15°C and a dominant wind direction of southwest monsoon. The annual precipitation is abundant, with a rainfall of more than 1000 mm. However, the dry and wet seasons are distinct, and the rainfall is concentrated from May to October, accounting for 84% of the total annual rainfall (Yuan et al., 2008). The predominant tree species are Pinus yunnanensis, Pinus armandii Franch., and Tsuga dumosa (D. Don) Eichler. The typical soil of the area is red soil.

The predominant tree species Pinus yunnanensis were surveyed within three circular sample plots (one plot in each peak). The radius of the corresponding circular sample plot is 18 meters (Malong Peak), 30 meters (Foding Peak) and 20 meters (Maer Peak). In each tree, the crown radius (CRi, m) was measured at a different height above crown top (CHi, m, i.e., the vertical height from crown top to each crown radius) along the crown profile; the diameter (DBH, cm) was measured at breast height (1.3 m aboveground), to the nearest 0.1 cm, and total tree height (TH, m) was measured to the nearest 0.1 m. The height to the base of the live crown (HCB, m), and the crown width (CW, m, the average values of two measures taken at eastwest direction and north–south direction) were also measured to the nearest 0.1 m in each tree (see Figure 1).

FIGURE 1
www.frontiersin.org

Figure 1 Tree crown measurement diagram of for Pinus yunnanensis.

CHi, crown height from treetop to each measurement point i (m); CRi, crown radius to each measurement point i (m).

The three data files were merged for extensive visual examination, screening, and outlier detection in an initial exploratory analysis before the pooled data were used for modeling. During the exploratory analysis, some obvious data errors were corrected. Tree crown measurements with relative crown height (RCH) than 1 were removed from the data set. Some trees were removed to ensure that each tree crown contained at least four crown radius measurements. Finally, the crown profile data used in this study were collected from 3,096 measured CR values of 516 trees from Pinus yunnanensis forests ranging in age from 16 to 45 years. A data summary is presented in Table 2.

TABLE 2
www.frontiersin.org

Table 2 Summary statistics of tree characteristic data for 516 sample trees.

Considering the multiple measurements for each sample tree crown, the relative crown heights from 0.1 to 1 m with an even interval of 0.1 m were selected to develop crown profile models. However, actual measurements of CH at each tree had non-equidistant space-steps. To overcome this difficulty, the crown profile data were used to obtain numerically interpolated values of CR using piecewise cubic Hermite interpolating polynomial (PCHIP) implemented in MATLAB (R2021b). This interpolation method was adopted because it had the characteristic of preserving shape well. Within each tree crown, CRh1CRh11denote CR in m at the 11 specified relative crown heights (h1 … h30) from crown top, ranging from 0.1 to 1 at an even interval of 0.1.

2.2 Multiple factors crown profile modeling based on LSTMs

Our research framework is illustrated in Figure 2. The whole framework comprises three parts: input features, model, and output. First, we need to extract the features we need from the data we have obtained. Then, we will divide the data into training and testing data, where the training data will be further split the validation data out. Next, the training and validation data are fed into the proposed LSTM and its variant algorithms for training. After the training process, we can put the testing data into the trained model, and it will output the predicted results.

FIGURE 2
www.frontiersin.org

Figure 2 The framework of multiple factors crown profile modeling based on deep learning.

2.2.1 Feature processing

2.2.1.1 Feature derivation

In order to extract the feature information of crown profile data to the maximum extent for LSTM algorithms to train the model, the feature polynomial fusion method and the feature correlation factor method were used to derive new features.

The feature polynomial fusion method can not only get the cross-term feature, but also get the higher order feature. The derivation of polynomial features combines low-dimensional features to obtain high-dimensional features, so that the LSTM models can capture the basic relationship of data to a greater extent and “learn” more data information. We selected the original data features CW, LCL and CH. After the third order polynomial addition, the number of features obtained was 19, of which the number of new features was 16. Feature derived expression were shown in equations (1) and (2).

Poly(a,b,c)n=(a,b,c,a2ab,ac,b2,bc,c2,,bcn1,cn)(1)
Numnew=k=1nk2+3k+223    (2)

Where Poly ()n is n-order feature derivation; a,b,c is three features of dataset; Numnew is the number of new derivation features.

For the feature correlation factor method, we defined the following composite tree factors: relative crown height (RCH= CH/LCL), tree slenderness coefficients (TSC= TH/DBH), crown length ratio (CLR= LCL/TH), crown fullness ratio (CFR=CW/LCL). These tree factors were also used to establish crown profile model.

2.2.1.2 Supervised transformation

Since crown profile is influenced by multiple environmental factors, the crown profile can be predicted by the change trend of the continuous adjacent space in each tree. Therefore, in order to fully consider the spatial sequence characteristics affecting the crown profile, CR data at a certain interval was taken as the new feature information, so that the model can fully learn the spatial factors and tree factors of the data. At the same time, the spatial sequence data of samples were combined into pairwise input and output formats, and the spatial sequence prediction problem was transformed into a supervised learning problem.

In this paper, CRi-1 (CR with lag space of 1) was taken as the spatial feature of position i, which together with the measured crown factors and the features derived from the features constituted the features of the model.

Finally, the modeled data has 28 input feature variables and one output variable, where the 28 input feature variables include 7 direct measurement features (DBH, TH, CW, HCW, HCB, LCL, CH), 16 polynomial features (CW*LCL, CW*CH, LCL2, LCL*CH, CH2, CW3, CW2*LCL, CW2*CH,LCL3, LCL2*CW,LCL2*CH, CH3, CH2*CW, CH2*LCL, CW*LCL*CH), 4 correlation factor features (RCH, TSC,CLR, CFR), and 1 supervised features of transformation (CRi-1).

2.2.1.3 Non-dimensionalize

Since the data involves multiple indicators that may affect the crown shape, the value range of each indicator is different. In order to unify the impact of index values on the model, this paper used the method of data normalization to make the sample data non-dimensionalize, and the normalized mathematical expression was equation (3).

x=xminmaxmin(3)

Where, x is the data value of an index before dimensionalization, max and min are respectively the maximum and minimum values of this index in all samples. x is the non-dimensionalized data value of this index.

2.2.2 LSTM algorithms and modeling scheme

LSTM is a modification of recurrent neural networks (RNNs) – neural networks that allow feedback loops to communicate data from a node in a forward layer to a node in a backward layer (Géron, 2022). LSTM networks overcome the problem of vanishing and exploding gradient problems by intelligently forgetting some past irrelevant information, and hence such network proves very suitable for modeling sequential data (Mehtab et al., 2020). LightGBM is a recent modification of the GB algorithm. It can outperform existing boosting frameworks on both efficiency and accuracy, with significantly lower memory consumption (Ke et al., 2017). JoZefowicz et al. (2015) analyzed the performance of more than 10,000 different LSTM permutations, some from the literatures (Krause et al., 2016; Chen et al., 2018; Kent and Salem, 2019; Lu et al., 2020) but most generated as LSTM variants, and found that some of the mutations did perform better than the classic LSTM, but not all, of the tasks, studied. In this study, the regression models based on the crown profile data sets were developed using LSTM and its variant algorithms, namely, stacked LSTMs-LightGBM, integrated LSTM-LightGBM, and hybrid LSTM-LightGBM. The models were optimized and extensively evaluated with the following modeling scheme. Vanila LSTM: Vanilla LSTM is defined as three parts: input layer, LSTM hidden layer, and fully connected output layer. We implemented a one-hidden-layer LSTM model using Keras with the Tensorflow backend (Figure 3A). The model hyperparameters were optimized within the predefined range: number of hidden units (100−1000), epoch (50−200), learning rate (0.0001,0.001,0.01), size of minibatch (10-100), optimization algorithm (Adam ()), activation function (relu for hidden layers), dropout rates for hidden layers (0, 0.1, 0.2).

FIGURE 3
www.frontiersin.org

Figure 3 The architecture of LSTM and its variant algorithms. (A) is Vanila LSTM algorithm; (B) is Stacked LSTMs-LightGBM algorithm; (C) is Integrated LSTM-LightGBM algorithm; (D) is Hybrid LSTM-LightGBM algorithm.

Stacked LSTMs-LightGBM: The proposed ensemble deep learning method consisted of three LSTMs and LightGBM was developed in this paper. More specifically, the LightGBM aggregated the outputs of different LSTMs. In the structure of stacked LSTMs-LightGBM, the predictions of a cluster of LSTMs were input into a LightGBM regression top-layer to produce the final forecasting (Figure 3B). The performance of ensemble learning depended vastly on the parameters of top-layer. Three hyperparameters governing the performance of the LightGBM regressor were optimized within the following number of leaves (“num_leaves”, 50−500), number of iterations (“n_estimators”,30−900), and learning rate (“learning_rate”, 10−4−10−1).

Integrated LSTM-LightGBM: Integrated LSTM-LightGBM was a combination of LSTM and LightGBM models to give full play to the advantages of combined model prediction. After the LSTM model was used to predict CR, LightGBM model was used to introduce the relaxation variable method to correct each predicted CR value for the calculated residual e (Figure 3C). The predicted value of LSTM model (CR) and the error corrected by LightGBM (e’) were integrated to obtain the prediction result of Integrated LSTM-LightGBM, namely (CR^ = CR + e’). LSTM parameters adopted the optimized parameters of the Vanila LSTM model, and LightGBM hyperparameter range was defined as follows: number of leaves (“num_leaves”, 10−200), number of iterations (“n_estimators”,30−300), and learning rate (“learning_rate”, 10−3−10−1).

Hybrid LSTM-LightGBM: The Hybrid LSTM-LightGBM took the predicted CR value of LSTM and LightGBM models as a feature input, and the measured CR value as an output. Ridge regression was performed, and the final hybrid model prediction result was obtained according to the obtained Ridge regression model (Figure 3D). The linear_model.RidgeCV function from scikit-learn was used to develop the regression models with hyperparameter optimized within the predefined ranges suggested: alphas (10−6−10).

The specific definitions of the relevant hyper-parameters are described in the Supplementary_Material.docx

2.3 Performance evaluation

The performance of the proposed algorithms was evaluated based on their balanced the determination coefficient (R2, equation (4)), the mean squared error (MSE, equation (5)), the mean absolute deviation (MAE, equation (6)), and the Mean Deviation (ME, equation (7)) values. The computation time of the proposed algorithms was also recorded to indicate the speed of each method.

R2=1i=1n(yiyi)2i=1n(yiyi¯)2(4)
MSE=1ni=1n(yiyi)2(5)
MAE=1ni=1n|yiyi|(6)
ME=i=0n(yiyi^)n (7)

where yi represents the observed value for the ith analytic tree ith; yi is the predicted value of ith observed value; n is the number of trees, yi¯ is the mean value for the observed.

The data split ratio adopted by the test was 8:1:1. That is, 80% of original data were used as the training set to train the model parameters, 10% of the original data was used as the validation set, which was used for model optimization during model training, and another 10% of original data was used as the test set to test the forecasting effect of model (Ju et al., 2019). All of models were run on a computer with Windows 11 operating system, Intel Core i7 CPU @ 3.20 GHz and RAM of 16.00 GB.

Understanding why a model makes a certain prediction can be as crucial as the prediction’s accuracy in the research of machine learning modeling. Here, we used SHAP (Lundberg and Lee 2017) to interpret the predictions of ensemble or deep learning models. We calculated SHAP value using shap package in Python. SHAP overall process: (1) Select the training set as the basic data set, and the explanation model instantiates an interpreter; (2) Select the test set as the explanation sample, and then calculate the SHAP value corresponding to each feature of each sample in the explanation sample through the interpreter. The greater the SHAP value, the greater the feature contribution.

3 Results

3.1 Prediction of the crown profile model

The parameters used in all the models were tuned by grid search or experiment, and the parameters of the ensemble learning model were the same as the parameters of the single model, which ensured the validity of ensemble learning. The optimizing parameters can ensure that all of parameters maintain the good performance in training and prediction. The optimized parameters of all the models used in the experiment are summarized in Table 3.

TABLE 3
www.frontiersin.org

Table 3 The parameter values for different models.

Figure 4 visualizes the training performance of Vanila LSTM. For Vanila LSTM model, the loss curves of the training and validation sets tended to be flat after parameter tuning, and the learning rate decreased to 0.0001. In the case of not amplifying the samples of the dataset, the model was fully trained at this time, but the effect of the training model had room for further improvement.

FIGURE 4
www.frontiersin.org

Figure 4 Accuracy and loss curves of training set and verification set (Vanila LSTM model).

The variant algorithm of LSTM has been used to further improve model accuracy. The training results of different crown profile models based on LSTM and its variant algorithms are listed Table 4. We observed that the LSTM’s variant algorithms performed higher accuracy compared to molecular LSTM algorithm (Vanila LSTM). From Table 4 and Figure 5, the proposed Hybrid LSTM-LightGBM performs better than two other variant algorithms with the minimum value of MAE as 0.2336, MSE as 0.1089 and ME as 3.91E-16 and the maximum value of R2 as 0.8656 for training datasets. And the better one of two other variant algorithms is Integrated LSTM-LightGBM with MAE as 0.2798, MSE as 0.1558, ME as -5.49E-12 and R2 as 0. 8078.For validation datasets, the proposed Integrated LSTM-LightGBM realizes the best rank, followed by Hybrid LSTM-LightGBM. Moreover, from the analysis of training time, the Hybrid LSTM-LightGBM required slightly more computational time compared to Integrated LSTM-LightGBM. We also observed that LightGBM was the fastest algorithm among the proposed algorithms as it consumed the shortest computation time for training datasets in this study, whereas Stacked LSTMs-LightGBM was the most time-consuming algorithm, requiring, on average, approximately 719 times longer computational time than LightGBM due to large number of algorithm parameters (Table 4).

TABLE 4
www.frontiersin.org

Table 4 The training results of different crown profile models based on LSTM and its variant algorithms.

FIGURE 5
www.frontiersin.org

Figure 5 Performance Evaluation for the five algorithms on the validation and test subsets.

Table 5 shows the error for different model predictions. As seen from the four forecasting performance indices (R2, ME, MAE and MSE) that the Integrated LSTM-LightGBM model had some degrees of advantages over LightGBM and Vanila LSTM, indicating that in the case of more complex data, LSTM’s variant algorithms were sufficient to learn crown profile features and it could predict accurately. Although the ensemble learning model (LightGBM) was not inferior to Vanila LSTM and its other variant algorithms in test results, the overall trend of the deep learning model was better. The comparable model performance on both validation and test sets indicated that the applied LSTM’s variant algorithms offered generalizability and transferability of the developed model to previously LSTM algorithm. Rank metric from testing results showed that Integrated LSTM-LightGBM performance achieved the best performance followed by LightGBM and Stacked LSTMs-LightGBM.

TABLE 5
www.frontiersin.org

Table 5 Comparison of test results of different crown profile models based on LSTM and its variant algorithms.

3.2 Relative importance of influential predictors

Considering that LSTM’s variant algorithms were composed of LightGBM and Vanila LSTM, the relative importance of potential influential predictors for predicting crown profile was investigated for each of LightGBM (Figure 6) and Vanila LSTM models (Figure 7). For beeswarm plot, each point is a result, its position on the x-axis represents the SHAP value of the feature, and the color represents the relative size of the feature, with red representing high and blue representing low.

FIGURE 6
www.frontiersin.org

Figure 6 Feature importance bar plot (mean (|SHAP value|), left plot) and beeswarm plot (mean value, right plot) for LightGBM model. Where color represents characteristic value (red high, blue low). * indicates a multiplication sign.

FIGURE 7
www.frontiersin.org

Figure 7 Feature importance bar plot (mean (|SHAP value|), left plot) and beeswarm plot (mean value, right plot) for Vanila LSTM model. Where color represents characteristic value (red high, blue low). * indicates a multiplication sign.

For LightGBM model, LAG(CR) (CRi-1), RCH and CW were found the most important predictors, while other variables were relatively insignificant variables (see Figure 6 feature importance bar plot). Furthermore, we found that LAG (CR), the most important feature, was basically positively correlated with CR (the model output); The RCH also had an obvious influence. Further analysis, if the RCH is large, the estimated value will be reduced significantly, because the leftmost point in the row of RCH is basically red in feature importance beeswarm plot; The SHAP value in the data point with large CW value is greater than 0, which belongs to positive correlation (see Figure 6 beeswarm plot).

For the Vanila LSTM model, the most important variables were found to be the Poly (CW, LCL, CH)3 and the original measurement tree factors (LAG(CR), DBH and TH), and the least important predictors were some derived factor (CLR, TSC, RCH, HCW, CFR and HCB), in that order (see Figure 7 left plot). The beeswarm plot (see Figure 7 right plot) shows that CH*LCL2, CW*LCL2, LCL3 and CW2*LCL are obviously positively correlated with predictors, while CH3 and CH2*CW are basically negatively correlated with predictors. The blue dots of CH*LCL2, CW*LCL2, LCL3 and CW3 variables are mainly concentrated in the area where SHAP value is less than 0, indicating that when their values are small, the estimate of CR will be reduced. For feature DBH, CH, LCL and CW, most of the points are diffuse in SHAP=0, so it has no effect on most of them, only a small part.

The relative importance of influential predictors was different for the LightGBM and Vanila LSTM models. There were only three factors that contribute significantly to LightGBM model, while almost all factors contributed to Vanila LSTM model. This phenomenon can explain why deep learning outperforms ensemble learning when there are many features.

4 Discussion

4.1 Performance and comparison of models

Previous studies have shown that machine learning has broad application in crown profile modeling. These machine learning includes MLP, SVR, RF, AdaBoost, GBDT and XGBoost (Tian et al., 2021; Chen et al., 2022). Among them, the ensemble learning algorithms can deal with complex nonlinear relationship and show strong prediction ability when predicting the crown profile (Chen et al., 2022).The purpose of this paper is to find the applicable model for crown profile prediction by comparing the ensemble and deep learning algorithms based on the same data format. So far, the deep learning algorithm has not been used to predict crown profile yet based on space-dependent data. Results from this study (see Table 4 and Table 5) showed that the Integrated LSTM-LightGBM model consistently obtained the best performance for crown profile prediction compared with LightGBM model. The overall trend of the deep learning model was better than the ensemble learning model.

An important finding was that there were only three feature variables that contribute significantly to LightGBM model, while almost all feature variables contributed to Vanila LSTM model (see Figures 6, 7). Two phenomena can be attributed to the finding. Firstly, the LSTM’s variant model was superior to the single LSTM model. The model relied on feature variables to develop its prediction algorithms. So, when the number of features contributed to model was less, the generalization ability of the whole model was therefore reduced and the accuracy of the predictor importance measurement was also affected. Secondly, the deep learning outperformed ensemble learning when there were many features (Zhang et al., 2018; Jan et al., 2019). The LSTM series was affected by more variables than LightGBM in this study. This finding supported the notion that more feature variables bring more room for improvement for LSTM series.

4.2 The interpretability of ensemble or deep learning models

Previous results have shown that the machine learning is an effective technique in improving the crown profile prediction accuracy. However, understanding why a model makes a certain prediction can be as crucial as the prediction’s accuracy (Lundberg and Lee, 2017). Interpretable machine learning techniques can generally be grouped into two categories: intrinsic interpretability and post-hoc interpretability (Du et al., 2019). Intrinsic interpretability incorporates interpretability directly to their structures, including decision tree, rule-based model, linear model, attention model, etc. In contrast, the post-hoc interpretability requires selecting and training a black-box model (ensemble or deep learning) and applying interpretability methods (feature importance, partial dependency graph) to explanate after training (Molnar, 2020). Current explanations are usually given in the format of feature importance vectors, which are a complete causal attribution explanation (Molnar et al., 2018). The explanation audiences, such as developers or researchers, can utilize the statistical analysis of the feature importance distribution to debug the models (Du et al., 2019).

In this paper, we adopt a unified framework for interpreting predictions of ensemble or deep learning models, SHAP. The relative importance of influential predictors was different for the LightGBM and Vanila LSTM models. There are more factors that contribute significantly to Vanila LSTM model compared to LightGBM model. The phenomenon can be attributed to two causes. Firstly, over-fitting can be a problem, especially for the LightGBM model. The generalization ability of the whole model was therefore reduced and the accuracy of the predictor importance measurement was also affected. Secondly, the problem of non-convex optimization makes convergence to a local optimum possible when the parameters of the model are learned and adjusted. This will lead to the deviation of the predictor importance’s estimation (Huang et al., 2019).

4.3 Limitations and further research

One of the limitations of this paper is that whether it is LightGBM or LSTM series, whether it is a Vanila LSTM model or an LSTM’s variant model, their robustness is difficult to guarantee for abnormal data and false data, but compared to the ensemble and deep learning models, the model generated by deep learning is still better, but it will consume more computing resources. Fortunately, the computational complexity of these algorithms is relatively small in crown profile modelling, and the main purpose of this paper is to explore the performance of deep learning methods. However, if we can solve the above problems, it will be more conducive to the improvement of the algorithm, this part of the task will be placed in our follow-up work.

Age has a certain influence on the crown profile (Yu et al., 2021). However, age variable inputs have not been used in this paper to predict crown profile because it is difficult to obtain the age of each tree for the uneven aged forest in this study. In future research, age variable may be used for improving the model predictions of crown profile. Moreover, current models may be biased when applied at a large scale as the crown profile is largely influenced by site quality, stand density, spatial structure (such as mingling, neighborhood comparison, and uniform angle index), and random variabilities caused by various stochastic factors that vary from stands to stands. Thus, the prediction bias can be reduced through the integration of all kinds of variability into the crown profile models (Sharma et al., 2017).

Several important model-agnostic interpretability methods such as Partial Dependence Plot (PDP), Individual Conditional Expectation (ICE), Permuted Feature Importance, Global Surrogate, Local Surrogate (LIME), SHAP Value exist, and while none of them are perfect, they can help researchers interpret the results of even very complex machine learning models (Zhou, 2019). For this study, we only used a method (SHAP Value) to interpret the predictions of ensemble or deep learning models. In the future work, multiple model-agnostic methods may be adopted, each of them represents a step toward more fully understanding machine learning models. As machine learning becomes more and more ubiquitous, grasping how these models find answers will be crucial to improving their performance and reliability.

5 Conclusions

This paper has introduced a novel method using deep learning prediction based on LSTMs, LightGBM and Ridge algorithm for crown profile forecasting. In the proposed LSTM’s variant models, a cluster of LSTMs with diverse hidden layers and neurons are employed separately to learn the information of crown profile. When compared with proposed prediction models including Vanila LSTM, LightGBM, Stacked LSTMs-LightGBM, Integrated LSTM-LightGBM and Hybrid LSTM-LightGBM, the pro-posed Hybrid LSTM-LightGBM can achieve a best forecasting performance with the minimum value of MAE, MSE and ME and the maximum value of R2 for training datasets, and the proposed Integrated LSTM-LightGBM can achieve a best forecasting performance on both validation and test sets. Furthermore, the analysis of feature importance of LightGBM and Vanila LSTM present that there are more factors that contribute significantly to Vanila LSTM model compared to LightGBM model. This phenomenon can explain why deep learning outperforms ensemble learning when there are more interrelated features. In conclusion, the following are the highlights of the study:

(1) LSTM’s variant algorithms obtained the best performance for crown profile prediction.

(2) The overall trend of the deep learning model was better than the ensemble learning model.

(3) We used a method (SHAP Value) to interpret the predictions of black-box models (ensemble or deep learning).

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Author contributions

YC and JW contributed to the study design and performed the formal analysis. YC performed the software analysis and wrote the first draft of the manuscript. JW contributed data curation. YC and JW contributed to the writing, review, and editing. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by the National Natural Science Foundation of China”Research on Structure Optimization of Pinus yunnanensis Secondary Forests Based on Reinforcement Learning” (grant number 32001313),Climate-sensitive Stand Biomass Model of Moso Bamboo (grant number 203402010801), Yunnan Fundamental Research Projects (grant number 202201AT070006) and Yunnan Postdoctoral Research Fund Projects (grant number ynbh20057).

Acknowledgments

We are grateful to Professor Wu Baoguo at Beijing Forestry University, for supplying administrative support.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2023.1093905/full#supplementary-material

References

Chen, Y., Dong, C., Wu, B. (2022). Crown profile modeling and prediction based on ensemble learning. Forests 13 (3), 410. doi: 10.3390/f13030410

CrossRef Full Text | Google Scholar

Chen, Q., Duan, G., Liu, Q., Ye, Q., Sharma, R. P., Chen, Y., et al. (2021). Estimating crown width in degraded forest: A two-level nonlinear mixed-effects crown width model for dacrydium pierrei and podocarpus imbricatus in tropical China. For. Ecol. Manage. 497, 119486. doi: 10.1016/j.foreco.2021.119486

CrossRef Full Text | Google Scholar

Chen, J., Zeng, G. Q., Zhou, W., Du, W., Lu, K. D. (2018). Wind speed forecasting using nonlinear-learning ensemble of deep learning time series prediction and extremal optimization. Energy Conversion Manage. 165, 681–695. doi: 10.1016/j.enconman.2018.03.098

CrossRef Full Text | Google Scholar

Crecente-Campo, F., Álvarez-González, J. G., Castedo-Dorado, F., Gómez-García, E., Diéguez-Aranda, U. (2013). Development of crown profile models for pinus pinaster ait. and pinus sylvestris l. @ in northwestern Spain. Forestry 86 (4), 481–491. doi: 10.1093/forestry/cpt019

CrossRef Full Text | Google Scholar

Crecente-Campo, F., Marshall, P., LeMay, V., Diéguez-Aranda, U. (2009). A crown profile model for pinus radiata d. don in northwestern Spain. For. Ecol. Manage. 257 (12), 2370–2379. doi: 10.1016/j.foreco.2009.03.038

CrossRef Full Text | Google Scholar

de-Miguel, S., Mehtätalo, L., Shater, Z., Kraid, B., Pukkala, T. (2012). Evaluating marginal and conditional predictions of taper models in the absence of calibration data. Can. J. For. Res. 42 (7), 1383–1394. doi: 10.1139/x2012-090

CrossRef Full Text | Google Scholar

Deng, W., Xiang, W., Ouyang, S., Hu, Y., Chen, L., Zeng, Y., et al. (2022). Spatially explicit optimization of the forest management tradeoff between timber production and carbon sequestration. Ecol. Indic. 142, 109193. doi: 10.1016/j.ecolind.2022.109193

CrossRef Full Text | Google Scholar

Di Salvatore, U., Marchi, M., Cantiani, P. (2021). Single-tree crown shape and crown volume models for pinus nigra JF Arnold in central Italy. Ann. For. Sci. 78 (3), 1–10. doi: 10.1007/s13595-021-01099-4

CrossRef Full Text | Google Scholar

Dong, S., Wang, P., Abbas, K. (2021). A survey on deep learning and its applications. Comput. Sci. Rev. 40, 100379. doi: 10.1016/j.cosrev.2021.100379

CrossRef Full Text | Google Scholar

Dong, C., Wu, B., Wang, C., Guo, Y., Han, Y. (2016). Study on crown profile models for Chinese fir (Cunninghamia lanceolata) in fujian province and its visualization simulation. Scandinavian J. For. Res. 31 (3), 302–313. doi: 10.1080/02827581.2015.1081982

CrossRef Full Text | Google Scholar

Du, M., Liu, N., Hu, X. (2019). Techniques for interpretable machine learning. Commun. ACM 63 (1), 68–77. doi: 10.1145/3359786

CrossRef Full Text | Google Scholar

Fawcett, D., Bennie, J., Anderson, K. (2021). Monitoring spring phenology of individual tree crowns using drone-acquired NDVI data. Remote Sens. Ecol. Conserv. 7 (2), 227–244. doi: 10.1002/rse2.184

CrossRef Full Text | Google Scholar

Fu, L., Sun, H., Sharma, R. P., Lei, Y., Zhang, H., Tang, S. (2013). Nonlinear mixed-effects crown width models for individual trees of Chinese fir (Cunninghamia lanceolata) in south-central China. For. Ecol. Manage. 302, 210–220. doi: 10.1016/j.foreco.2013.03.036

CrossRef Full Text | Google Scholar

Fu, L., Zhang, H., Sharma, R. P., Pang, L., Wang, G. (2017). A generalized nonlinear mixed-effects height to crown base model for Mongolian oak in northeast China. For. Ecol. Manage. 384, 34–43. doi: 10.1016/j.foreco.2016.09.012

CrossRef Full Text | Google Scholar

Géron, A. (2022). Hands-on machine learning with scikit-learn, keras, and TensorFlow (O'Reilly Media, Inc).

Google Scholar

Hann, D. W. (1999). An adjustable predictor of crown profile for stand-grown Douglas-fir trees. For. Sci. 45 (2), 217–225. doi: 10.1093/forestscience/45.2.217

CrossRef Full Text | Google Scholar

Huang, X., Gao, L., Crosbie, R. S., Zhang, N., Fu, G., Doble, R. (2019). Groundwater recharge prediction using linear regression, multi-layer perception network, and deep learning. Water 11 (9), 1879. doi: 10.3390/w11091879

CrossRef Full Text | Google Scholar

Jan, B., Farman, H., Khan, M., Imran, M., Islam, I. U., Ahmad, A., et al. (2019). Deep learning in big data analytics: a comparative study. Comput. Electrical Eng. 75, 275–287. doi: 10.1016/j.compeleceng.2017.12.009

CrossRef Full Text | Google Scholar

Jia, W., Chen, D. (2019). Nonlinear mixed-effects height to crown base and crown length dynamic models using the branch mortality technique for a Korean larch (Larix olgensis) plantation in northeast China. J. Forestry Res. 30 (6), 2095–2109. doi: 10.1007/s11676-019-01045-1

CrossRef Full Text | Google Scholar

Jin, Z. Z., Peng, J. (2004). Yunnan pine (Pinus yunnanensis franch.) (Kunming: Yunnan Science and Technology Press), 1–66.

Google Scholar

Jozefowicz, R., Zaremba, W., Sutskever, I. (2015). “An empirical exploration of recurrent network architectures,” in Proceedings of the International Conference on Machine Learning, Lille, France. 6–11.

Google Scholar

Ju, Y., Sun, G., Chen, Q., Zhang, M., Zhu, H., Rehman, M. U. (2019). A model combining convolutional neural network and LightGBM algorithm for ultra-short-term wind power forecasting. IEEE Access 7, 28309–28318. doi: 10.1109/ACCESS.2019.2901920

CrossRef Full Text | Google Scholar

Ke, G., Meng, Q., Finley, T., Wang, T., Chen, W., Ma, W., et al. (2017). “Lightgbm: A highly efficient gradient boosting decision tree,” in Advances in neural information processing systems, vol. 30.

Google Scholar

Kent, D., Salem, F. (2019). “Performance of three slim variants of the long short-term memory (LSTM) layer,” in 2019 IEEE 62nd International midwest Symposium on circuits and Systems (MWSCAS). 307–310.

Google Scholar

Krause, B., Lu, L., Murray, I., Renals, S. (2016). Multiplicative LSTM for sequence modelling. arXiv 1609, 07959. doi: 10.48550/arXiv.1609.07959

CrossRef Full Text | Google Scholar

Lejeune, G., Ung, C. H., Fortin, M., Guo, X. J., Lambert, M. C., Ruel, J. C. (2009). A simple stem taper model with mixed effects for boreal black spruce. Eur. J. For. Res. 128 (5), 505–513. doi: 10.1007/s10342-009-0300-8

CrossRef Full Text | Google Scholar

Lu, Z., Lv, W., Cao, Y., Xie, Z., Peng, H., Du, B. (2020). LSTM variants meet graph neural networks for road speed prediction. Neurocomputing 400, 34–45. doi: 10.1016/j.neucom.2020.03.031

CrossRef Full Text | Google Scholar

Lundberg, S. M., Lee, S. I. (2017). “A unified approach to interpreting model predictions,” in Advances in neural information processing systems, vol. 30.

Google Scholar

McCulloch, C. E., Searle, S. R. (2001). “Generalized, linear, and mixed models,” in Wiley Series in probability and statistics (NJ, USA: Wiley: Hoboken).

Google Scholar

Mehtab, S., Sen, J., Dutta, A. (2020). “Stock price prediction using machine learning and LSTM-based deep learning models,” in Symposium on machine learning and metaheuristics algorithms, and applications (Singapore: Springer), 88–106.

Google Scholar

Molnar, C. (2020) Interpretable machine learning. Available at: https://www.Lulu.com.

Google Scholar

Molnar, C., Casalicchio, G., Bischl, B. (2018). Iml: An r package for interpretable machine learning. J. Open Source Software 3 (26), 786. doi: 10.21105/joss.00786

CrossRef Full Text | Google Scholar

Myers, R. H. (1990). Classical and modern regression with applications. 2nd ed (Belmont: Duxbury Press).

Google Scholar

Pleșoianu, A. I., Stupariu, M. S., Șandric, I., Pătru-Stupariu, I., Drăguț, L. (2020). Individual tree-crown detection and species classification in very high-resolution remote sensing imagery using a deep learning ensemble model. Remote Sens. 12 (15), 2426. doi: 10.3390/rs12152426

CrossRef Full Text | Google Scholar

Rautiainen, M., Mõttus, M., Stenberg, P., Ervasti, S. (2008). Crown envelope shape measurements and models. Silva Fennica 42 (1), 19. doi: 10.14214/sf.261

CrossRef Full Text | Google Scholar

Reichstein, M., Camps-Valls, G., Stevens, B., Jung, M., Denzler, J., Carvalhais, N. (2019). Deep learning and process understanding for data-driven earth system science. Nature 566 (7743), 195–204. doi: 10.1038/s41586-019-0912-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Sharma, R. P., Bílek, L., Vacek, Z., Vacek, S. (2017). Modelling crown width–diameter relationship for scots pine in the central Europe. Trees 31 (6), 1875–1889. doi: 10.1007/s00468-017-1593-8

CrossRef Full Text | Google Scholar

Sharma, R. P., Vacek, Z., Vacek, S. (2016). Individual tree crown width models for Norway spruce and European beech in Czech republic. For. Ecol. Manage. 366, 208–220. doi: 10.1016/j.foreco.2016.01.040

CrossRef Full Text | Google Scholar

Singh, A., Thakur, N., Sharma, A. (2016). “A review of supervised machine learning algorithms,” in 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom). 1310–1315.

Google Scholar

Sun, Y., Gao, H., Li, F. (2017). Using linear mixed-effects models with quantile regression to simulate the crown profile of planted pinus sylvestris var. mongolica trees. Forests 8 (11), 446. doi: 10.3390/f8110446

CrossRef Full Text | Google Scholar

Sun, C., Huang, C., Zhang, H., Chen, B., An, F., Wang, L., et al. (2022). Individual tree crown segmentation and crown width extraction from a heightmap derived from aerial laser scanning data using a deep learning framework. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.914974

CrossRef Full Text | Google Scholar

Sun, H. L., Li, S. C., Xiong, W. L., Yang, Z. R., Cui, B. S. (2008). Influence of slope on root system anchorage of Pinus yunnanensis. Ecol. Eng. 32 (1), 60–67. doi: 10.1016/j.ecoleng.2007.09.002

CrossRef Full Text | Google Scholar

Tian, Y., Wu, B., Su, X., Qi, Y., Chen, Y., Min, Z. (2021). A crown contour envelope model of Chinese fir based on random forest and mathematical modeling. Forests 12 (1), 48. doi: 10.3390/f12010048

CrossRef Full Text | Google Scholar

Wang, W., Ge, F., Hou, Z., Meng, J. (2021). Predicting crown width and length using nonlinear mixed-effects models: a test of competition measures using Chinese fir (Cunninghamia lanceolata (Lamb.) hook.). Ann. For. Sci. 78 (3), 1–17. doi: 10.1007/s13595-021-01092-x

CrossRef Full Text | Google Scholar

Wang, M., Kane, M. B., Zhao, D. (2017). Correlation-regression analysis for understanding dominant height projection accuracy. For. Sci. 63 (6), 549–558.

Google Scholar

Wang, C., Wu, B., Chen, Y., Qi, Y. (2019). Development of crown profile models for Chinese fir using non-linear mixed-effects modelling. Nat. Environ. Pollut. Technol. 18 (4), 1349–1361.

Google Scholar

Wu, Z. Y. (1986). Flora yunnanica (Tomus 4): Spermatophyta (Beijing: Science Press), 54–57.

Google Scholar

Wu, M. (2019). Python Deep learning: based on PyTorch (Beijing: China Machine Press).

Google Scholar

Xu, Y., Woeste, K., Cai, N., Kang, X., Li, G., Chen, S., et al. (2016). Variation in needle and cone traits in natural populations of Pinus yunnanensis. J. forestry Res. 27 (1), 41–49. doi: 10.1007/s11676-015-0153-6

CrossRef Full Text | Google Scholar

Yang, M., Mou, Y., Liu, S., Meng, Y., Liu, Z., Li, P., et al. (2022). Detecting and mapping tree crowns based on convolutional neural network and Google earth images. Int. J. Appl. Earth Observation Geoinformation 108, 102764. doi: 10.1016/j.jag.2022.102764

CrossRef Full Text | Google Scholar

Yuan, R., Yang, S., Wang, B. (2008). Study on the altitudinal pattern of vegetation distribution along the eastern slope of cangshan Mountain,Yunnan,China. J. Yunnan Univ. (Natural Sci. Edition). 03, 318–325.

Google Scholar

Yu, P., Gao, R., Zhang, D., Liu, Z. P. (2021). Predicting coastal algal blooms with environmental factors by machine learning methods. Ecol. Indic. 123, 107334. doi: 10.1016/j.ecolind.2020.107334

CrossRef Full Text | Google Scholar

Zhang, Q., Yang, L. T., Chen, Z., Li, P. (2018). A survey on deep learning for big data. Inf. Fusion 42, 146–157. doi: 10.1016/j.inffus.2017.10.006

CrossRef Full Text | Google Scholar

Zhou, X. (2019). Interpretability methods in machine learning: A brief survey (Two Sigm). Available at: https://www.twosigma.com/articles/interpretability-methods-in-machine-learning-a-brief-survey/.

Google Scholar

Keywords: crown profile, LSTM and its variant algorithms, LightGBM, SHAP, Pinus yunnanensis

Citation: Chen Y and Wang J (2023) Deep learning for crown profile modelling of Pinus yunnanensis secondary forests in Southwest China. Front. Plant Sci. 14:1093905. doi: 10.3389/fpls.2023.1093905

Received: 09 November 2022; Accepted: 02 January 2023;
Published: 03 February 2023.

Edited by:

Veronique Letort, CentraleSupelec, Paris-Saclay University, France

Reviewed by:

Jiangchuan Fan, Beijing Research Center for Information Technology in Agriculture, China
Rabia Saleem, University of Derby, United Kingdom

Copyright © 2023 Chen and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jianming Wang, wjm@dali.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.