Next Article in Journal
Nitrogen Addition Alleviates Cadmium Toxicity in Eleocarpus glabripetalus Seedlings
Previous Article in Journal
Morphological Crown Attributes Contributing to Differential Height and Diameter Growth among Families in a Coastal Douglas-Fir Progeny Test: Crown Structural Evidence for Crop Ideotypes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Species Classification Using Optimized Features Derived from Light Detection and Ranging Point Clouds Based on Fractal Geometry and Quantitative Structure Model

1
Key Laboratory of Mine Environmental Monitoring and Improving around Poyang Lake of Ministry of Natural Resources, East China University of Technology, Nanchang 330013, China
2
School of Surveying and Geoinformation Engineering, East China University of Technology, Nanchang 330013, China
3
Powerchina Guiyang Engineering Corporation Limited, Guiyang 550081, China
*
Author to whom correspondence should be addressed.
Submission received: 23 April 2023 / Revised: 2 June 2023 / Accepted: 16 June 2023 / Published: 19 June 2023
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)

Abstract

:
Tree species classification is a ubiquitous task in the forest inventory field. Only directly measured feature vectors have been applied to most existing methods that use LiDAR technology for tree species classification. As a result, it is difficult to obtain a satisfactory tree species classification performance. To solve this challenge, the authors of this paper developed two new kinds of feature vectors, including fractal geometry-based feature vectors and quantitative structural model (QSM)-based feature vectors. In terms of fractal geometry, both two fractal parameters were extracted as feature vectors for reflecting how tree architecture is distributed in three-dimensional space. In terms of QSM, the ratio of length change and the ratio of radius change of different branches were extracted as feature vectors. To reduce the feature vector dimensionality and explore valuable feature vectors, feature vector dimension reduction was conducted using the classification and regression tree (CART). Five hundred and sixty-eight individual trees with five tree species were selected for evaluating the performance of the developed feature vectors. The experimental results indicate that the tree species of Fagus sylvatica achieved the highest overall accuracy, which is 98.06%, while Quercus petraea obtained the lowest overall accuracy, which is 96.65%. Four other classical supervised learning methods were adopted for comparison. The comparison result indicates that the proposed method outperformed the other four supervised learning methods no matter which accuracy indicator was adopted. In comparison with the relevant method, the eight feature vectors developed in this paper also performed much better. This indicates that the fractal geometry-based feature vectors and QSM-based feature vectors developed in this paper can effectively improve the performance of tree species classification.

1. Introduction

Forests are one of the most important ecosystems on earth and are of great significance to economic construction, climate change and human survival [1,2,3]. The composition of forest tree species has a direct impact on ecological attributes, such as the greenhouse gas absorption capacity and the resource utilization rate. The richness of forest tree species can also reflect the level of forest productivity and forest biodiversity [4,5]. Therefore, tree species classification has become a ubiquitous task in forest resource inventory, and also, a popular issue in current research [4,6,7].
LiDAR is an active remote sensing technology. The laser pulses emitted by an LiDAR system can not only obtain the three-dimensional coordinate information about a target, but also penetrate vegetation and accurately depict spatial structure information about the vegetation canopy [8,9,10]. In recent years, with the rapid development of LiDAR technology, its sampling rate and sampling accuracy have obviously been improved, making LiDAR technology widely used in the forest inventory field [11].
Numerous tree species identification methods using LiDAR technology have been proposed in recent years. Lin and Herold [12] proposed explicit tree structure (ETS) feature parameters derived from using terrestrial laser scanning (TLS) to classify boreal tree species. In their method, there are five types of ETS feature parameter, including the structural characteristics of an entire tree, stem, branches, crown and leaves. And then, leave-one-out-cross-validation was adopted in SVM to evaluate the performance of extracted ETS feature parameters. The experimental results showed that the average and maximum classification accuracies for 40 samples of four typical boreal tree species were up to 77.5% and 90%, respectively. In the method proposed by Åkerblom et al. [6], they first used the TreeQSM method proposed by Raumonen et al. [13] to construct a quantitative structure model (QSM) of tree point clouds. Fifteen types of features were extracted from the QSM, and then optimized based on the performance of different feature sets in their method. The experimental results showed that using ten optimized features can achieve a maximum classification accuracy of 96.9%, which proved that geometric features extracted from QSM can effectively achieve tree species classification. Terryn et al. [2] proposed two new feature parameters, including the branch angle ratio and relative volume, on the basis of the study proposed by Åkerblom et al. [6]. Seven hundred and fifty-eight trees of five species (Acer pseudoplatanus, Fraxinus excelsior, Crataegus monogyna, Corylus avellana and Quercus robur) in a 1.4-hectare mixed deciduous forest site were classified. They achieved about 80% overall accuracy for the experimental sample after using principal component analysis on the feature parameters. Xi et al. [4] extracted 32 classification features in total after the QSM was constructed using the method proposed by Xi et al. [14]. In their experiments, the classification performances of seven deep learning and six machine learning classifiers were compared. Among them, the PointNet++ method achieved the best overall classification accuracy, and compared to other deep learning methods, the training time of this method was also the shortest. However, their experimental results also demonstrated that when there are a few tree samples, this method is prone to overfitting, which will reduce the classification accuracy. Liu et al. [15] constructed a deep neural network learning model named LayerNet for the species classification of LiDAR tree point clouds in simple forest areas. They validated the classification performance of this method using both airborne and ground-based LiDAR datasets. The experimental results showed that their proposed method can achieve overall classification accuracies of 88.8% and 92.5% for airborne and ground-based datasets, respectively. Their method can also achieve more satisfactory accuracy and Kappa coefficients in comparison to those of traditional machine learning and deep learning methods. However, the information redundancy caused by large feature dimensions will affect the performance of this method.
In addition to the tree species classification methods using LiDAR point clouds, some researchers also tried to fuse LiDAR data and other data sources to obtain tree species classification results. Dalpont et al. [16] investigated the effect of combining hyperspectral data with LiDAR data for classification in complex forest environments with more than 19 tree species. Their study showed that combining hyperspectral data with LiDAR data could improve the tree species identification performance in cases where the spectral information were relatively similar, but the tree heights obviously differed between species. Kim et al. [17] further investigated the tree species identification performance using multiple LiDAR intensity data. They used intensity information derived from leaf-on and leaf-off point clouds of trees in the same forest site to identify coniferous and broadleaved species. The experimental results showed that when leaf-off point clouds were used, the classification accuracy was higher than that when leaf-on point clouds were used, and the highest accuracy of 90.06% was obtained when the combination of both leaf-on and leaf-off point clouds were used. Puttonen et al. [18] combined hyperspectral data and tree shape features from LiDAR to form a fused dataset. And then, the fused dataset was used to classify 24 tree samples of three tree species (Birch, Norway spruce and Scots pine) in the SVM. The classification performances using the fused datasets and two single source datasets (hyperspectral data and tree shape features) were compared. The experimental results showed that the best classification performance was realized when the fused dataset was used, which was able to achieve an accuracy above 85%. Othmani et al. [19] classified tree species using the three-dimensional texture feature of bark. In their method, a two-dimensional image representing the three-dimensional geometric texture feature of the bark at around breast height was generated first. Next, the multiresolution analysis technique was applied to the image to extract the texture feature. The extracted features can achieve a classification accuracy up to 86.93% for experimental tree samples, but this method had struggles to recognize Pine and Hornbeam species. Zhang et al. [20] proposed a method that combined canopy information derived from LiDAR and spectral information in hyperspectral imagery to classify urban tree species. In their method, the canopy height model derived from LiDAR data was segmented via object-based image analysis to obtain individual tree crowns. Then, significant bands of individual tree spectrums were selected via minimum noise fraction transformation, and the classification performance of the selected bands was measured using random forest and multi class classifiers.
Although fusing LiDAR point clouds with other remote sensing data, such as hyperspecrtal image, can improve the performance of tree species classification, the fusion process is prone to error. Realizing the high-precision fusion of multi-source data is still a challenge. Thus, the classification of different tree species using LiDAR point clouds alone is still conducted by researchers. However, recognizing tree species from LiDAR point clouds still suffers from the following two problems. One is how to explore more effective features for tree species classification. The other one is how to obtain a combination of the feature vectors to achieve a higher classification accuracy, while reducing the feature vector dimension. To solve these challenges, this paper proposes a tree species classification method based on the combination of developed fractal geometry-based features and QSM-based features. In this method, three kinds of feature vectors were explored, including directly measured feature vectors, fractal geometry-based feature vectors and QSM-based feature vectors. In this paper, fifteen feature vectors in total were calculated. To reduce feature vector dimension, the classification and regression tree (CART) was adopted to analyze the importance of each feature vector. Among the fifteen feature vectors, ten important feature vectors were extracted for further analysis. Thereafter, to further reduce feature vector dimension, eight feature vectors with a high accuracy level and high occurrence frequency were extracted. The support vector machine (SVM) classification method was applied for tree species classification using the final eight extracted feature vectors. The experimental results show that the proposed method can obtain satisfying tree species classification results.

2. Materials and Methods

2.1. Datasets

This paper adopted the individual tree point clouds provided by Weiser et al. [21] for evaluating the performance of tree species classification. In this dataset, twelve forest plots of approximately 1 ha were scanned using laser scanners of different platforms, including airborne LiDAR, unscrewed aerial vehicle (UAV)-based LiDAR and terrestrial LiDAR. The twelve forest plots are located in mixed central European forests close to Bretten and Karlsruhe, in the federal state of Baden-Württemberg, Germany. This dataset provided segmented individual tree point clouds from different platforms, corresponding tree species information and tree metrics. Since the feature vectors developed in this paper need to construct the QSM, only the individual tree point clouds of high quality obtained via UAV-based LiDAR and terrestrial LiDAR were selected for testing. Moreover, to balance the number of samples of different tree species, some tree species with smaller number of trees were ignored. Eventually, 568 individual trees of five tree species, such as Fagus sylvatica, Picea abies, Pinus sylvestris, Pseudotsuga menziesii and Quercus petraea, were selected for evaluating the performance of the developed feature vectors. The detailed information for the selected individual trees is tabulated in Table 1. Figure 1 shows several individual trees of the five tree species. It can be found that all these individual trees own high-quality point clouds. Thus, the QSM of each individual tree can be accurately built. Meanwhile, several tree metrics, such as DBH, can be precisely calculated.
Figure 2a,b shows the distribution of tree heights and DBHs of all the individual trees tabulated in Table 1. It can be found that the height of trees varies greatly from species to species. Combined with Table 1, the average of height of different tree species changes from 20.84 m to 36.30 m. In terms of DBH, the average values of DBH of these five species do not change greatly. From Table 1, it can also be found that the smallest average value of DBH is 0.19 m, while the largest one is 0.35 m.

2.2. Method

The flowchart of the proposed method is shown in Figure 3. In the proposed method, multi-dimensional feature vectors were first extracted from each individual tree points. And then, the dimensionality reduction of feature vectors was applied for reducing the computation burden. Subsequently, the final eight feature vectors were obtained based on the importance and occurrence frequency of each feature vector. Lastly, support vector machines (SVM) were adopted to obtain the final tree species classification results. Three main steps are involved in this paper, namely: (i) multi-dimensional feature vectors extraction, (ii) the dimensionality reduction of feature vectors, and (iii) the selection of feature vectors combination.

2.2.1. Multi-Dimensional Feature Vectors Extraction

In this method, three kinds of feature vectors were developed, including directly measured feature vectors, fractal geometry-based feature vectors and QSM-based feature vectors. These three kinds of feature vectors are shown in Figure 4.
(i)Directly measured feature vectors
The first kind of feature vectors can be directly measured from the individual tree point cloud, including the tree height ( H ), diameter at breast height (DBH), ratio of DBH and H ( κ ), major axis of canopy ellipse ( l 1 ), minor axis of canopy ellipse ( l 2 ), eccentricity of canopy ellipse ( δ ), canopy convex hull area ( S 1 ), tree point cloud projection area ( S 2 ) and entropy ( E ). H is calculated as the elevation difference between the highest and lowest points within one individual tree. DBH is the tree diameter at 1.3 m from the tree’s root. In this paper, DBH was calculated using the method proposed by Di Wang et al. [22]. l 1 and l 2 are calculated as the major and minor axes of the fitted ellipse for canopy points. δ is the corresponding eccentricity of the fitted ellipse. S 1 is calculated as the convex hull area of canopy point. S 2 is the horizontal projection area of all tree points, which can be calculated by gridding all the points with x and y coordinates. E can be calculated via voxelizing the tree points. The ratio of the number of points within each voxel to the number of all tree points can be calculated as p i . E is defined as Equation (1):
E = i = 1 n p i × log ( p i )
where n is the number of voxels.
(ii) Fractal geometry-based feature vectors
Fractal geometry considers that many objects have the hierarchy of self-similarity, which can be observed at different scales. When the geometry is appropriately scaled up or down, the whole structural feature does not change. Guzman Q et al. [23] have proven that there is a noticeable correlation between the fractal geometry parameters and tree metrics, such as tree height, DBH and crown area. Since these tree metrics can be used for tree species identification, the authors of this paper tried to apply geometry parameters for classifying different tree species.
The fractal geometry parameters can be calculated based on the box counting method. This method considers that the individual tree points can be covered by a series of boxes, as shown in Figure 5a. When the voxel size changes from large to small, the number of voxels for covering the tree points will be distinctly increased. A log–log linear regression model can be built using the voxel size and the number of voxels, as defined in Equation (2):
log N = d M B × log 1 V + I n t e r c e p t M B
As shown in Figure 5b, d M B is the slope of the linear regression equation, which is defined as the fractal dimension. I n t e r c e p t M B is the intercept of the linear model, which is defined as the fractal intercept. Both these two fractal parameters were used for tree species classification in this paper.
(iii) QSM-based feature vectors
The QSM can reflect the structure features and spatial topology of an individual tree. Thus, the authors of this paper tried to extract feature vectors based on the QSM for tree species classification. In this paper, the QSM was constructed using the TreeQSM method proposed by Raumonen et al. [13]. In TreeQSM, the model of an individual tree is constructed using a series of fitted cylinders. Moreover, the topology of each tree branch can also be built via TreeQSM. That is, TreeQSM can separate branches into different levels. In general, different tree species have different ratios of length and diameter of branches at different levels, such as level 1, level 2, level 3, etc., as shown in Figure 6. Thus, this paper developed four feature vectors for tree species classification, as defined in Equations (3)–(6):
R 12 = i = 1 K ( R 2 i R 1 i ) / K
R 23 = j = 1 M ( R 3 j R 2 j ) / M
L 12 = i = 1 K ( L 2 i L 1 ) / K
L 23 = j = 1 M ( L 3 j L 2 j ) / M
where R 12 and R 23 ; L 12 and L 23 represent the ratio of radius and the ratio of length of branches at different levels, respectively.

2.2.2. Dimensionality Reduction of Feature Vectors

As mentioned above, these three kinds of feature vectors contain fifteen feature vectors in total. Although fifteen feature vectors were developed in this paper, not all the feature vectors have noticeable contributions to tree species identification. To reduce feature vector dimensionality, this paper further analyzes the relative importance of each feature vector using the classification and regression tree (CART). The CART can be used to judge the importance of each feature vector in terms of classification and further eliminate the relatively low level of importance of classification features to achieve dimensionality reduction (Zhouxin Xi et al. [4]). In the CART method, the Gini index was adopted for selecting partition attribute when the decision tree was built, which is defined as Equation (7).
G i n i ( D ) = 1 k = 1 N p k 2
where pk is the proportion of class k. Gini (D) reflects the probability that two samples are randomly selected from the dataset, D , and their categories are inconsistent. Obviously, the smaller G i n i ( D ) is, the more purity the dataset D is. When building the CART decision tree, the leaf node corresponds to the decision result, while the branch nodes correspond to an attribute being split. In this paper, the branch nodes represent attributes of these feature vectors for splitting. Ideally, when a decision tree is built, the samples contained by the branch nodes should belong to the same category as far as possible. This means that the purity of the nodes increases. On the contrary, at each node, the risk of splitting is estimated to be the node impurity. The authors of this paper calculated the importance of each feature vector by summing changes in the risk due to splits on each node.
The relative importance of each vector calculated using the CART is shown in Figure 7. It can be found that although there are nine feature vectors involved in the directly measured feature vectors, three of them show a relatively low level of importance for tree species identification. These three feature vectors are DBH, the ratio ( κ ) of DBH and H and the eccentricity ( δ ) of canopy ellipse. In terms of fractal geometry-based feature vectors, I n t e r c e p t M B is more important than d M B is. From Figure 7, it is easy to see that all the four QSM-based feature vectors are functional in classifying tree species. Comparatively speaking, the relative importance of L 12 is lower than that of the other three QSM-based feature vectors. To reduce the feature vector dimension, five feature vectors with a low level of relative importance in these three kinds of feature vectors were ignored in the following SVM classification. These five feature vectors are κ , δ and DBH in the directly measured feature vectors, d M B in the fractal geometry-based feature vectors and L 12 in the QSM-based feature vectors.

2.2.3. Selection of Feature Vectors Combination

After dimensionality reduction, ten feature vectors were retained for further processing. To select the final combination of feature vectors, 4-fold cross validation was conducted via applying different combination of feature vectors with different dimensions. The 4-fold cross validation method is shown in Figure 8. The datasets were classified into four sets. Three sets were selected for training, while the remaining one was used for testing. The mean accuracy for the four results is given as the classification accuracies.
The results of using different combination of feature vectors are tabulated in Table 2. It can be found that the dimensionality of the feature vectors changes from 1 to 10. Different dimensionality of the feature vectors contains a different number of combination of feature vectors. For instance, when the dimensionality is 1, there are 10 different combinations, such as { H }, { l 1 } and { L 23 }, etc. Although the dimensionality is same, different combinations of feature vectors have different classification accuracies. From Table 2, it is easy to find that when the dimensionality is 3, the lowest overall accuracy of the combination of feature vectors is 0.421, while the highest overall accuracy is 0.845. In total, there are 1023 different combinations of feature vectors. In general, if one feature vector is noticeably effective at tree species classifying, the occurrence frequency of this feature vector should be high among the combinations with a high classifying accuracy. Thus, to further analysis the relative importance of each feature vector, the occurrence frequency of each feature vector among the different combinations of feature vectors whose classifying accuracy is higher than 0.85 was statistically counted.

3. Experimental Result and Analysis

3.1. Feature Vector Dimensional Reduction

As mentioned above, to further analyze the relative importance of each feature vector, the occurrence frequency of each feature vector was statistically counted. The statistical result is shown in Figure 9. It can be found that the occurrence frequency of tree height (H) feature vector is the highest. Comparatively speaking, the minor axis of the canopy ellipse ( l 2 ) and canopy convex hull area ( S 1 ) have the lowest frequency. These two feature vectors were ignored in the final SVM classification. Thus, eight feature vectors were actually selected for tree species classification. The reason for this can be explained as follows.
As can be seen in Table 2, when the feature vector dimension is less than eight, its lowest or highest overall accuracies are all smaller than those of the feature sets whose dimension is eight. In addition, when the feature vector dimension is larger than eight, its highest overall accuracy is still smaller than that of the feature sets whose dimension is eight. Moreover, a smaller feature vector dimension means that fewer calculations are involved. Thus, the final eight feature vectors, including H, R 23 , L 23 , I n t e r c e p t M B , R 12 , E , S 2 and l 1 were adopted for tree species classification in this paper.

3.2. Experimental Results

Five accuracy indexes were adopted for evaluating the performance of the proposed method, including the overall accuracy (OA), precision (Pre), recall (Rec), F1 score ( F 1 ) and kappa coefficient (Ka). These five accuracy indexes are defined as Equations (8)–(13), respectively:
O A = T P + T N T P + T N + F P + F N
P r e = T P T P + F P
R e c = T P T P + F N
F 1 = 2 × P r e × R e c P r e + R e c
K a = O A P e 1 P e
P e = ( T P + F N ) × ( T P + F P ) + ( F P + T N ) × ( F N + T N ) N 2
where T P is the number of trees correctly classified as the specific species, T N is the number of trees correctly classified as other species, F N is the number of trees misclassified as other species and F P is number of trees misclassified as the specific species by other species. The confusion matrix of the classification results in this paper is shown in Figure 10.
According to the classification results shown in Figure 10, the five accuracy indexes were calculated, as tabulated in Table 3. In terms of O A , all the tree species have a higher classification overall accuracy, whose O A are all high than 95%. In terms of P r e , the precision values of all the tree species are all larger than 90%. This means that the proposed method can obtain satisfactory classification results for all the five tree species. In Figure 10, it can be seen that there are 134 trees whose tree species are fagus sylvatica. In the classification results, 126 trees of fagus sylvatica were correctly classified. In terms of other accuracy indexes, such as F 1 and K a , the proposed method can all obtain satisfactory classification results. Thus, it can be concluded that the developed feature vectors have a remarkable effect on tree species classification.
In addition to SVM, there are many other classical supervised learning methods, such as Adaptive Boosting (AdaBoost), K-Nearest Neighbors (KNN), Naive Bayes (NB) and random forest (RF). The authors of this paper further compared the accuracy metrics of tree species classification via SVM with the ones of other supervised learning methods. The comparison results are shown in Table 4. In Table 4, it can be seen that the proposed method can obtain the best tree species classification results, no matter which accuracy index is adopted. In terms of O A , the proposed method achieved 93.31% overall accuracy, while the overall accuracies of the other four methods are all smaller than 90%. In terms of F 1 and K a , the proposed method still performed the best. Thus, it can be concluded that SVM adopted in this paper outperformed the other four classical supervised learning methods in terms of tree species classification.

4. Discussion

In this paper, the dimensionality of feature vectors was reduced from fifteen to eight. In achieving this, the computation burden can be relieved. The final feature vectors combination contains H, R 23 , L 23 , I n t e r c e p t M B , R 12 , E , S 2 and l 1 , according to the occurrence frequency of each feature vector. As mentioned above, this paper provides a strategy for feature dimension reduction. To further discuss the performance of the reduced feature vectors, the authors of this paper tested all the tree species classification results using all the feature vector combinations when the feature dimension was eight. The results are shown in Table 5. In Table 5, it can be seen that there are forty-five different feature vector combinations in total. In terms of O A , there are only four combinations whose O A is a little larger than the feature vector combination adopted in this paper was. Four combinations are colored in blue, as shown in Table 5. It is easy to see that their O A s are less than one percentage points higher than that of the proposed eight feature vectors. Meanwhile, the performances of the proposed eight feature vectors are better than the ones of all the other forty-feature vector combinations, no matter which accuracy indicator was adopted. Thus, it can be concluded that using the occurrence frequency of each feature vector will provide an effective dimension reduction method for tree species classification, while maintaining a promising tree species classification performance.
Akerblom et al. [6] developed feature vectors using the QSM for recognizing tree species. In their method, three different types of feature vectors, namely the stem branch, crown and tree, were presented. There were fifteen feature vectors in total, including the stem branch angle, stem branch cluster size, three stem branches’ radius, stem branch length, stem branch distance, crown start height, crown height, crown evenness, crown diameter/height, DBH/height ratio, DBH/tree volume, DBH/minimum tree radius, volume below 55% of the total height, cylinder length/tree volume and shedding ratio. Three different classification methods, namely KNN, multinomial regression (MNR) and SVM (including three different kernel functions, such as linear, polynomial and radial basis functions, which are SVMlin, SVMpol and SVMrbf, respectively) were applied for tree species classification. The authors of this paper compared their performance with that of the proposed method. The comparison is shown in Table 6. In Table 6, it can be seen that compared with the fifteen feature vectors developed by Akerblom et al. [6], although only eight feature vectors were adopted in this paper, the proposed method obtained much better tree species classification results. Thus, it can be concluded that the eight feature vectors developed in this paper are more effective at tree species classification.

5. Conclusions

Tree species have a direct impact on forests’ productivity and diversity. To accurately and efficiently identify tree species, this paper proposed a tree species classification method based on the combination of fractal geometric feature vectors and QSM feature vectors. In this paper, three different types of feature vectors were first extracted. Specially, fractal geometry-based feature vectors, including fractal dimension and intercept, were developed. Meanwhile, four different QSM feature vectors were also introduced to obtain better tree species classification results. Successively, the feature vector dimensionality was reduced by analyzing the relative importance for tree species identification using the CART method. To further conduct feature dimension reduction, the occurrence frequency of each feature vector was statistically calculated using different combinations of feature vectors. Eventually, SVM was applied to obtain tree species classification results using the reduced eight feature vectors. Five hundred and sixty-eight individual tree point clouds with five tree species were used for testing. The experimental results show that the developed feature vectors are effective at identifying different tree species. The overall accuracies for the five tree species are all greater than 95%. In the comparison with the other four classical supervised learning methods, the proposed method still performs the best. Compared with the relevant method, the eight feature vectors developed in this paper also performed much better. This indicates that the fractal geometry-based feature vectors and QSM-based feature vectors developed in this paper can effectively improve the performance of tree species classification.

Author Contributions

Conceptualization, Z.H.; experimental analysis, Z.C. and P.C.; writing—original draft preparation, Z.H. and Z.C.; writing—review and editing, Y.X. and P.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (NSF) (42161060, 41801325, 41861052, 42174055 and 41962018), Research on Key technologies of spatio-temporal information and collaborative design in engineering (YJZD2020-01), the China Post-Doctoral Science Foundation (2019M661858), the Natural Science Foundation of Jiangxi Province (20192BAB217010) and East China University of Technology’s Ph.D. Project (DHBK2017155) for their financial support.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pfeifer, N.; Gorte, B.; Winterhalder, D. Automatic Reconstruction of Single Trees from Terrestrial Laser Scanner Data. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2004, 35, 114–119. [Google Scholar]
  2. Terryn, L.; Calders, K.; Disney, M.; Origo, N.; Malhi, Y.; Newnham, G.; Raumonen, P.; Å Kerblom, M.; Verbeeck, H. Tree species classification using structural features derived from terrestrial laser scanning. ISPRS J. Photogramm. 2020, 168, 170–181. [Google Scholar] [CrossRef]
  3. Knapp, N.; Fischer, R.; Cazcarra-Bes, V.; Huth, A. Structure metrics to generalize biomass estimation from lidar across forest types from different continents. Remote Sens. Environ. 2020, 237, 111597. [Google Scholar] [CrossRef]
  4. Xi, Z.; Hopkinson, C.; Rood, S.B.; Peddle, D.R. See the forest and the trees: Effective machine and deep learning algorithms for wood filtering and tree species classification from terrestrial laser scanning. ISPRS J. Photogramm. 2020, 168, 1–16. [Google Scholar] [CrossRef]
  5. Pretzsch, H.; Biber, P. Tree species mixing can increase maximum stand density. Can. J. For. Res. 2016, 46, 1179–1193. [Google Scholar] [CrossRef] [Green Version]
  6. Åkerblom, M.; Raumonen, P.; Mäkipää, R.; Kaasalainen, M. Automatic tree species recognition with quantitative structure models. Remote Sens. Environ. 2017, 191, 1–12. [Google Scholar] [CrossRef]
  7. Ghosh, A.; Fassnacht, F.E.; Joshi, P.K.; Koch, B. A framework for mapping tree species combining hyperspectral and LiDAR data: Role of selected classifiers and sensor across three spatial scales. Int. J. Appl. Earth Obs. 2014, 26, 49–63. [Google Scholar] [CrossRef]
  8. Jin, S.; Sun, X.; Wu, F.; Su, Y.; Li, Y.; Song, S.; Xu, K.; Ma, Q.; Baret, F.; Jiang, D.; et al. Lidar sheds new light on plant phenomics for plant breeding and management: Recent advances and future prospects. ISPRS J. Photogramm. 2021, 171, 202–223. [Google Scholar] [CrossRef]
  9. Hui, Z.; Jin, S.; Xia, Y.; Wang, L.; Yevenyo Ziggah, Y.; Cheng, P. Wood and leaf separation from terrestrial LiDAR point clouds based on mode points evolution. ISPRS J. Photogramm. 2021, 178, 219–239. [Google Scholar] [CrossRef]
  10. Hui, Z.; Jin, S.; Li, D.; Ziggah, Y.Y.; Liu, B. Individual Tree Extraction from Terrestrial LiDAR Point Clouds Based on Transfer Learning and Gaussian Mixture Model Separation. Remote Sens. 2021, 13, 223. [Google Scholar] [CrossRef]
  11. Guo, Q.; Su, Y.; Hu, T.; Guan, H.; Jin, S.; Zhang, J.; Zhao, X.; Xu, K.; Wei, D.; Kelly, M. Lidar Boosts 3D Ecological Observations and Modeling:A Review and Perspective. IEEE Geosci. Remote Sens. 2021, 9, 232–257. [Google Scholar] [CrossRef]
  12. Lin, Y.; Herold, M. Tree species classification based on explicit tree structure feature parameters derived from static terrestrial laser scanning data. Agr. For. Meteorol. 2016, 216, 105–114. [Google Scholar] [CrossRef]
  13. Raumonen, P.; Kaasalainen, M.; Åkerblom, M.; Kaasalainen, S.; Kaartinen, H.; Vastaranta, M.; Holopainen, M.; Disney, M.; Lewis, P. Fast Automatic Precision Tree Models from Terrestrial Laser Scanner Data. Remote Sens. 2013, 5, 491–520. [Google Scholar] [CrossRef] [Green Version]
  14. Xi, Z.; Hopkinson, C.; Chasmer, L. Filtering Stems and Branches from Terrestrial Laser Scanning Point Clouds Using Deep 3-D Fully Convolutional Networks. Remote Sens. 2018, 10, 1215. [Google Scholar] [CrossRef] [Green Version]
  15. Liu, M.; Han, Z.; Chen, Y.; Liu, Z.; Han, Y. Tree species classification of LiDAR data based on 3D deep learning. Measurement 2021, 177, 109301. [Google Scholar] [CrossRef]
  16. Dalponte, M.; Bruzzone, L.; Gianelle, D. Fusion of Hyperspectral and LIDAR Remote Sensing Data for Classification of Complex Forest Areas. IEEE Trans. Geosci. Remote 2008, 46, 1416–1427. [Google Scholar] [CrossRef] [Green Version]
  17. Kim, S.; McGaughey, R.J.; Andersen, H.; Schreuder, G. Tree species differentiation using intensity data derived from leaf-on and leaf-off airborne laser scanner data. Remote Sens. Environ. 2009, 113, 1575–1586. [Google Scholar] [CrossRef]
  18. Puttonen, E.; Suomalainen, J.; Hakala, T.; Räikkönen, E.; Kaartinen, H.; Kaasalainen, S.; Litkey, P. Tree species classification from fused active hyperspectral reflectance and LIDAR measurements. For. Ecol. Manag. 2010, 260, 1843–1852. [Google Scholar] [CrossRef]
  19. Othmani, A.; Lew Yan Voon, L.F.C.; Stolz, C.; Piboule, A. Single tree species classification from Terrestrial Laser Scanning data for forest inventory. Pattern Recogn. Lett. 2013, 34, 2144–2150. [Google Scholar] [CrossRef]
  20. Zhang, Z.; Kazakova, A.; Moskal, L.; Styers, D. Object-Based Tree Species Classification in Urban Ecosystems Using LiDAR and Hyperspectral Data. Forests 2016, 7, 122. [Google Scholar] [CrossRef] [Green Version]
  21. Weiser, H.; Schäfer, J.; Winiwarter, L.; Krašovec, N.; Fassnacht, F.E.; Höfle, B. Individual tree point clouds and tree measurements from multi-platform laser scanning in German forests. Earth Syst. Sci. Data 2022, 14, 2989–3012. [Google Scholar] [CrossRef]
  22. Wang, D.; Kankare, V.; Puttonen, E.; Hollaus, M.; Pfeifer, N. Reconstructing Stem Cross Section Shapes from Terrestrial Laser Scanning. IEEE Geosci. Remote Sens. 2017, 14, 272–276. [Google Scholar] [CrossRef]
  23. Guzmán, Q.J.A.; Sharp, I.; Alencastro, F.; Sánchez Azofeifa, G.A. On the relationship of fractal geometry and tree–stand metrics on point clouds derived from terrestrial laser scanning. Methods Ecol. Evol. 2020, 11, 1309–1318. [Google Scholar] [CrossRef]
Figure 1. Samples of individual trees of different tree species. (ac) are samples of tree species of FagSyl; (df) are samples of tree species of PicAbi; (gi) are samples of tree species of PinSyl; (jl) are samples of tree species of PseMen and (mo) are samples of tree species of QuePet.
Figure 1. Samples of individual trees of different tree species. (ac) are samples of tree species of FagSyl; (df) are samples of tree species of PicAbi; (gi) are samples of tree species of PinSyl; (jl) are samples of tree species of PseMen and (mo) are samples of tree species of QuePet.
Forests 14 01265 g001aForests 14 01265 g001b
Figure 2. Distribution of tree heights and DBHs of the five tree species: (a) tree height distribution of different tree species; (b) DBH distribution of different tree species. In the figure, the white dotted line represents median, while the two black dotted lines represent 25 percentile and 75 percentile, respectively.
Figure 2. Distribution of tree heights and DBHs of the five tree species: (a) tree height distribution of different tree species; (b) DBH distribution of different tree species. In the figure, the white dotted line represents median, while the two black dotted lines represent 25 percentile and 75 percentile, respectively.
Forests 14 01265 g002
Figure 3. Flowchart of the proposed method.
Figure 3. Flowchart of the proposed method.
Forests 14 01265 g003
Figure 4. Multi-dimensional feature vectors extraction.
Figure 4. Multi-dimensional feature vectors extraction.
Forests 14 01265 g004
Figure 5. Fractal geometry parameters calculation using the box counting method: (a) tree points covered with a series of boxes; (b) a log–log linear regression using voxel size ( V ) and the number of voxels ( N ).
Figure 5. Fractal geometry parameters calculation using the box counting method: (a) tree points covered with a series of boxes; (b) a log–log linear regression using voxel size ( V ) and the number of voxels ( N ).
Forests 14 01265 g005
Figure 6. The sketch map of branches of different levels in QSM.
Figure 6. The sketch map of branches of different levels in QSM.
Forests 14 01265 g006
Figure 7. Relative importance of each vector calculated using the CART.
Figure 7. Relative importance of each vector calculated using the CART.
Forests 14 01265 g007
Figure 8. Fourfold cross validation.
Figure 8. Fourfold cross validation.
Forests 14 01265 g008
Figure 9. Occurrence frequency of each feature vector.
Figure 9. Occurrence frequency of each feature vector.
Forests 14 01265 g009
Figure 10. Confusion matrix of the classification results in this paper.
Figure 10. Confusion matrix of the classification results in this paper.
Forests 14 01265 g010
Table 1. Information about selected individual trees. Std represents the standard deviation. abbr means abbreviation.
Table 1. Information about selected individual trees. Std represents the standard deviation. abbr means abbreviation.
SpeciesNumber of TreesAverage of Height (m)Std of Tree Height (m)Average of DBH (m)Std of DBH (m)
Fagus sylvatica
(abbr: FagSyl)
12928.024.080.350.15
Picea abies
(abbr: PicAbi)
12320.845.590.260.12
Pinus sylvestris
(abbr: PinSyl)
8129.203.450.290.13
Pseudotsuga menziesii
(abbr: PseMen)
12436.304.760.280.12
Quercus petraea
(abbr: QuePet)
11121.187.460.190.08
Table 2. Overall accuracy of different combination of feature vectors.
Table 2. Overall accuracy of different combination of feature vectors.
DimensionalityLowest
Overall
Accuracy
Highest
Overall
Accuracy
Combination of Feature VectorsNumber of Combinations
10.3540.583{ H },{ l 1 }…{ L 23 }10
20.4260.734{ H , l 1 },{ H , l 2 }…{ R 23 , L 23 }45
30.4210.845{ H , l 1 , l 2 },{ H , l 1 , S 1 }…{ R 12 , R 23 , L 23 }120
40.4400.896{ H , l 1 , l 2 , S 1 },{ H , l 1 , l 2 , S 2 }…
{ I n t e r c e p t M B , R 12 , R 23 , L 23 }
210
50.5210.938{ H , l 1 , l 2 , S 1 , S 2 },{ H , l 1 , l 2 , S 1 , E }…
{ E , I n t e r c e p t M B , R 12 , R 23 , L 23 }
252
60.6100.928{ H , l 1 , l 2 , S 1 , S 2 , E },
{ H , l 1 , l 2 , S 1 , S 2 , I n t e r c e p t M B }…
{ S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }
210
70.6880.931{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B },
{ H , l 1 , l 2 , S 1 , S 2 , E , R 12 }…
{ S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }
120
80.7830.947{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 },
{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 23 }…
{ l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }
45
90.8800.938{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 },
{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , L 23 }...
{ l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }
10
100.9230.923{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }1
In total///1023
Table 3. Accuracy metrics calculation results for the five different tree species.
Table 3. Accuracy metrics calculation results for the five different tree species.
FagSylPicAbiPinSylPseMenQuePet
O A (%)98.0696.8397.7197.3696.65
P r e (%)94.0392.0092.5095.0492.59
R e c (%)97.6793.5091.3692.7490.09
F 1 (%)95.8292.7491.9393.8891.32
K a (%)94.5690.7290.5992.1989.25
Table 4. Accuracy metrics of different supervised learning methods.
Table 4. Accuracy metrics of different supervised learning methods.
AdaBoostKNNNBRFThe Proposed Method
O A (%)88.0389.6181.3489.7993.31
P r e (%)87.6989.6381.1189.6693.23
R e c (%)87.3489.2181.3788.9593.07
F 1 (%)87.4989.3780.9589.2293.14
K a (%)85.3687.2677.6087.4491.72
Table 5. Accuracy metrics comparison of different combination of feature vectors. (The background color of the combination of feature vectors adopted in this paper was red, while the background color of four feature vector combinations, whose performances are a little better than the proposed feature vector combinations was blue.)
Table 5. Accuracy metrics comparison of different combination of feature vectors. (The background color of the combination of feature vectors adopted in this paper was red, while the background color of four feature vector combinations, whose performances are a little better than the proposed feature vector combinations was blue.)
CombinationOA (%)Pre (%)Rec (%)F1 (%)Ka (%)
{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B ,R12}85.7484.9985.1985.0582.08
{ H , l 1 , l 2 , S 1 , S 2 ,, I n t e r c e p t M B , R 23 }88.7387.8388.0187.985.85
{ H , l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , L 23 }82.0480.7981.1880.8277.44
{ H , l 1 , l 2 , S 1 , S 2 , E , R 12 , R 23 }85.7484.7184.6384.6682.06
{ H , l 1 , l 2 , S 1 , S 2 , E , R 12 , L 23 }84.8684.2284.6484.3980.98
{ H , l 1 , l 2 , S 1 , S 2 , E , R 23 , L 23 }89.6189.4589.2989.3486.93
{ H , l 1 , l 2 , S 1 , S 2 , I n t e r c e p t M B , R 12 , R 23 }88.3887.6287.6587.6285.39
{, l 1 , l 2 , S 1 , S 2 , I n t e r c e p t M B , R 12 , L 23 }88.3887.9888.2688.0985.4
{ H , l 1 , l 2 , S 1 , S 2 , I n t e r c e p t M B , R 23 , L 23 }91.0290.8590.6790.7588.71
{ H , l 1 , l 2 , S 1 , S 2 , R 12 , R 23 , L 23 }90.3290.1689.9190.0087.82
{ H , l 1 , l 2 , S 1 , E , I n t e r c e p t M B , R 12 , R 23 }87.8586.8686.8586.8484.73
{ H , l 1 , l 2 , S 1 , E , I n t e r c e p t M B , R 12 ,L23}87.6887.3987.2587.2684.5
{ H , l 1 , l 2 , S 1 , E , I n t e r c e p t M B , R 23 , L 23 }92.7892.9392.3392.5890.91
{ H , l 1 , l 2 , S 1 , E , R 12 , R 23 , L 23 }93.1392.8592.8692.8591.37
{ H , l 1 ,, S 1 , I n t e r c e p t M B , R 12 , R 23 , L 23 }91.991.6191.6891.6489.82
{ H , l 1 , l 2 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 }89.4488.6588.6188.6186.72
{ H , l 1 , l 2 , S 2 , E , I n t e r c e p t M B , R 12 , L 23 }87.8587.6587.3787.4684.72
{ H , l 1 , l 2 , S 2 , E , I n t e r c e p t M B , R 23 , L 23 }93.1393.4492.8293.0891.36
{ H , l 1 , l 2 , S 2 , E , R 12 , R 23 , L 23 }91.0290.5490.890.6588.72
{ H , l 1 , l 2 , S 2 , I n t e r c e p t M B , R 12 , R 23 , L 23 }92.4392.2492.2892.2390.48
{ H , l 1 , l 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }92.2592.0892.192.0790.26
{ H ,l1, S 1 , S 2 ,E, I n t e r c e p t M B , R 12 , R 23 }88.5687.9887.8687.8985.61
{ H , l 1 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , L 23 }89.7989.3789.5889.4687.17
{ H , l 1 , S 1 , S 2 , E , I n t e r c e p t M B , R 23 , L 23 } 94.01 94.23 93.8993.9692.47
{ H , l 1 , S 1 , S 2 , E , R 12 , R 23 , L 23 }91.3791.0391.1491.0589.16
{ H , l 1 , S 1 , S 2 , I n t e r c e p t M B , R 12 , R 23 , L 23 }90.6790.4890.3290.3888.26
{ H , l 1 , S 1 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }94.1994.1293.9494.0292.69
{ H , l 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }93.3193.2393.0793.1491.72
{ H , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 }89.6188.9589.1389.0386.95
{ H , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , L 23 }90.1490.1189.7289.8687.6
{ H , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 23 , L 23 }91.991.9391.6191.7489.81
{ H , l 2 , S 1 , S 2 , E , R 12 , R 23 , L 23 }90.6790.3990.4990.3988.27
{ H , l 2 , S 1 , S 2 , I n t e r c e p t M B , R 12 , R 23 , L 23 }91.991.791.7891.6589.83
{ H , l 2 , S 1 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }93.8493.8193.7593.7892.25
{ H , l 2 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }93.6693.3993.3393.3592.03
{ H , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }91.3791.0791.2791.1489.16
{ l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 }81.5181.0981.0581.0476.76
{ l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , L 23 }82.9282.7782.5382.6278.52
{ l 1 , l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 23 , L 23 }85.2185.8885.6685.7181.4
{ l 1 , l 2 , S 1 , S 2 , E , R 12 , R 23 , L 23 }80.4680.9980.6980.875.42
{ l 1 , l 2 , S 1 , S 2 , I n t e r c e p t M B , R 12 , R 23 , L 23 }85.2186.185.9185.9981.4
{ l 1 , l 2 , S 1 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }88.0388.3788.1688.2384.94
{ l 1 , l 2 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }86.2786.9286.786.7582.73
{ l 1 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }86.2786.7286.7486.6782.74
{ l 2 , S 1 , S 2 , E , I n t e r c e p t M B , R 12 , R 23 , L 23 }86.2786.7386.6486.6282.73
Table 6. Accuracy metrics comparison with the method proposed by Åkerblom et al. [6].
Table 6. Accuracy metrics comparison with the method proposed by Åkerblom et al. [6].
Åkerblom et al. [6]The Proposed Method
KNNMNRSVMlinSVMpolSVMrbf
O A (%)79.7581.8781.6975.8882.0493.31
P r e (%)80.0080.9480.8778.7981.8293.23
R e c (%)78.1880.9380.9673.5480.7593.07
F 1 (%)78.4480.9080.9174.0181.0993.14
K a (%)74.4677.2176.9969.4777.3791.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hui, Z.; Cai, Z.; Xu, P.; Xia, Y.; Cheng, P. Tree Species Classification Using Optimized Features Derived from Light Detection and Ranging Point Clouds Based on Fractal Geometry and Quantitative Structure Model. Forests 2023, 14, 1265. https://0-doi-org.brum.beds.ac.uk/10.3390/f14061265

AMA Style

Hui Z, Cai Z, Xu P, Xia Y, Cheng P. Tree Species Classification Using Optimized Features Derived from Light Detection and Ranging Point Clouds Based on Fractal Geometry and Quantitative Structure Model. Forests. 2023; 14(6):1265. https://0-doi-org.brum.beds.ac.uk/10.3390/f14061265

Chicago/Turabian Style

Hui, Zhenyang, Zhaochen Cai, Peng Xu, Yuanping Xia, and Penggen Cheng. 2023. "Tree Species Classification Using Optimized Features Derived from Light Detection and Ranging Point Clouds Based on Fractal Geometry and Quantitative Structure Model" Forests 14, no. 6: 1265. https://0-doi-org.brum.beds.ac.uk/10.3390/f14061265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop