Next Article in Journal
Use of Harvester Data to Estimate the Amount of Merchantable Non-Utilized Woody Material Remaining after Mechanized Cut-to-Length Forest Operations
Previous Article in Journal
China’s Forest Eco-Bank Project: An Analysis Based on the Actor-Network Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Verification of a Deep Learning-Based Tree Species Identification Model Using Images of Broadleaf and Coniferous Tree Leaves

1
Graduate School of Life and Environmental Sciences, Kyoto Prefectural University, Kyoto 6068522, Japan
2
Faculty of Life and Environmental Sciences, Kyoto Prefectural University, Kyoto 6068522, Japan
*
Author to whom correspondence should be addressed.
Submission received: 12 May 2022 / Revised: 15 June 2022 / Accepted: 15 June 2022 / Published: 16 June 2022

Abstract

:
The objective of this study was to verify the accuracy of tree species identification using deep learning with leaf images of broadleaf and coniferous trees in outdoor photographs. For each of 12 broadleaf and eight coniferous tree species, we acquired 300 photographs of leaves and used those to produce 72,000 256 × 256-pixel images. We used Caffe as the deep learning framework and AlexNet and GoogLeNet as the deep learning algorithms. We constructed four learning models that combined two learning patterns: one for individual classification of 20 species and the other for two-group classification (broadleaf vs. coniferous trees), with and without data augmentation, respectively. The performance of the proposed model was evaluated according to the MCC and F-score. Both classification models exhibited very high accuracy for all learning patterns; the highest MCC was 0.997 for GoogLeNet with data augmentation. The classification accuracy was higher for broadleaf trees when the model was trained using broadleaf only; for coniferous trees, the classification accuracy was higher when the model was trained using both tree types simultaneously than when it was trained using coniferous trees only.

1. Introduction

We developed an auto-tree-identification system based on leaf images for mobile terminals [1,2,3,4,5,6]. In previous studies, we focused mainly on woody plant species and investigated whether tree species could be identified from leaf images. In tree species identification using leaf images, images are captured in a manner such that only a single leaf is included. Pre-processing (e.g., image processing) is then performed. Various information from the images is used as identification criteria; this information includes shape features [7,8,9,10,11], texture [12,13] and leaf venation [14,15,16]. Supervised learning with machine learning methods [17], such as neural networks and decision trees, is used to discriminate among candidate species [18]. Minowa et al. [1] did not achieve high classification accuracy when they applied machine learning to shape features. Minowa et al. [2] demonstrated that classification accuracy could be improved by combining shape and leaf venation information. However, while classification accuracy was high for training data, performance was poor for test data. Several of our subsequent studies [4,5,6] demonstrated that deep learning can be applied to tree species identification with high classification accuracy, even for test data.
Deep learning refers to a neural network model that consists of multiple layers, such as convolutional neural networks (CNNs); such models have recently achieved excellent results with respect to image recognition. Prior to the emergence of deep learning, most image recognition methods generally used pre-extracted image features. The accuracy of image recognition is affected by the type of features used. However, image features are difficult to specify because they vary among objects. Thus, the extraction of image features is greatly influenced by researcher and developer experience [19]. Deep learning does not require users to define image features in advance. The user inputs a large amount of training data into the computer, which then extracts image features from the training data to perform recognition. Most deep learning applications are easy to use, publicly available and free of charge, which facilitates their use by researchers in various fields; thus, people in fields other than machine learning can conduct research using deep learning. With regard to tree species identification using leaf images, deep learning enables the omission of complicated image processing performed on the leaves for analysis. Accordingly, the analyst is not required to process numerous leaf images using image analysis software to extract shape features and venation information from a single leaf image.
State-of-the-art deep learning algorithms have been proposed at large-scale international competitions for fields related to image recognition. These include ILSVRC [20], COCO Challenge [21], ImageCLEF/LifeCLEF [22] and Places2 Challenge [23]. Among these competitions, LifeCLEF includes a sub-competition called PlantCLEF, in which participants strive to achieve high classification accuracy for various types of plant datasets [17]. In the 2015 PlantCLEF competition, plants were divided into seven parts (branch, entire plant, leaf [photographic image], leaf [scan or scan-like image], flower, fruit and stem) to show competitive classification accuracy. The 2015 PlantCLEF competition included 113,205 images of 1000 species of herbs, trees and ferns, mostly from France. The GoogLeNet model had the highest overall classification accuracy, but its mean reciprocal rank was only 0.667 (i.e., not high). Ghazi et al. [17] applied transfer learning [24] and other techniques to the AlexNet, GoogLeNet and VGGNet deep learning models, using the same dataset as the 2015 PlantCLEF competition; they achieved high classification accuracies with leaf scans and scan-like images but lower classification accuracies with leaf photographs. Because ImageCLEF and PlantCLEF focus on competition with deep learning algorithms, they do not extensively explore the tendency to misclassify targets. We speculated that the low classification accuracy of the 2015 PlantCLEF was partly related to the large number of plant species used for identification. Deep learning has been used to assess the extent of plant diseases [25,26], identify tree species based on Lidar and drone-acquired images [27,28,29] and to distinguish broadleaf and conifer species [30,31,32,33,34]. Most such studies use scanned images of single leaves rather than multiple leaves or images from databases such as PlantCLEF2012 [35] or the Swedish leaf dataset [36], whereas few studies have directly used images of multiple leaves.
In our previous studies, tree species were broadly classified into broadleaf [5] and coniferous [6] leaf images; deep learning was applied to tree species identification. The main objective of those studies was to determine whether deep learning could be used to identify tree species based on photographs of multiple leaves that were acquired outdoors. First, we conducted tree species identification for broadleaf trees using machine learning [5]. Second, we conducted tree species identification for coniferous trees based on the results of broadleaf tree analyses [6]. Therefore, our constructed deep learning model could not be applied to datasets that contain both broadleaf and coniferous trees. However, on the basis of tree species identification using machine learning and deep learning, we aim to develop a system that can perform tree species identification for numerous users using various mobile terminals in the future. For users who are familiar with tree species identification, it is appropriate to construct a tree species identification system that is adjusted for broadleaf and coniferous trees in advance. For users who are unfamiliar with tree species, the species of a tree should be automatically identified by simply acquiring a photograph of its leaves.
The objective of this study was to verify the accuracy of tree species identification using deep learning with broadleaf and coniferous leaf images acquired in the field (except for instances in which broadleaf and coniferous images are simultaneously reflected in a single image). We used 12 broadleaf and eight coniferous trees. Thus, we calculated the accuracy of tree species identification when 20 species were divided into two groups: all 12 broadleaf trees were in the “broadleaf group” and all eight coniferous trees were in the “coniferous group”. We then compared the initial accuracy with the accuracy of tree species identification when only broadleaf trees [5] and only coniferous trees [6] were used. Some deep learning models, such as pipelined and linked models, have functions that allow multiple deep learning models to be connected and regarded as a single model [37,38]. The size (capacity) of a learning model is especially important when the model is implemented in a mobile terminal, but the above functions enable construction of a model that is much smaller than conventional models. Moreover, if separately trained models can be combined, broadleaf and coniferous trees can be classified in advance and then identified using their respectively trained models. Overall, we focused on determining the accuracy of tree species identification when all tree species are classified simultaneously (i.e., without distinguishing between broadleaf and coniferous trees), compared to the accuracy when two groups of trees are classified separately.

2. Materials and Methods

2.1. Study Sites

Leaf images of 12 broadleaf tree species [5] and 8 coniferous tree species [6] were used for analysis at the Kyoto Prefectural University campus and Kyoto Botanical Gardens (Figure 1). All broadleaf tree species had simple leaves but were classified into three different categories: simple leaves with smooth margins, simple leaves with toothed margins and lobed leaves. We photographed 300 leaves from each tree. Previous studies of broadleaf trees [2,3,4,5] used these same leaf shapes, which facilitated comparison with our results.

2.2. Photographic Methods

Because broadleaf and coniferous trees were photographed by different photographers on different dates, a NICON COOLPIX A900 [39] was used for broadleaf trees and a Galaxy S9 was used for coniferous trees [40]. Photographs were acquired from September 2020 to November 2020. When possible, photographs were acquired on sunny days under calm conditions to ensure consistent image color tone. The distance between the subject and the camera was approximately 0.3–1.0 m. All images were in color. We did not process additional (i.e., non-leaf) elements, such as the sky, ground or buildings, which were present in the background of some photographs. We acquired photographs from various angles to avoid capturing the same leaf when possible; to ensure that the leaves were generally centered, we did not enlarge images. In deep learning, background objects are important. Since the distance between the subject and the camera was approximately 0.3–1.0 m, in many cases the background is barely visible in the image in this study. If the background objects are reflected in the image, we acquired leaf images that were taken at various angles so that the background of the leaf image would be various background images, such as the sky, ground (soil, concrete), buildings and other structures.

2.3. Image Processing and Data Augmentation

We processed the photographs using ImageJ 1.50 open-source image processing software [41]. Image sizes were 256 × 256 pixels. The following procedure was used for data augmentation of the training and test data. Data augmentation is a method for efficiently increasing the amount of learning data in deep learning. While deep learning does not require the definition of features, it requires the preparation of large volumes of training data. Because this is laborious, several methods have been proposed for efficient deep learning with less data. Typical methods include data augmentation and transfer learning [19,24,42].
First, we obtained three leaf images from each photograph by slightly varying the cropping window (Process 1 in Figure 2). Next, we rotated each cropped image clockwise by 90, 180 or 270 degrees (Process 2 in Figure 2). These procedures yielded 12 images per photograph and 3600 images per species; the final dataset comprised 72,000 images.

2.4. CNN Algorithms

Of the available CNN algorithms, we chose AlexNet and GoogLeNet, which are representative algorithms from previous studies [5,6]. AlexNet, which won the ILSVRC 2012 competition, is a deep learning model that consists of five convolutional layers, three max-pooling layers, and three fully connected layers; it thus comprises 11 layers [43]. GoogLeNet, which won the ILSVRC 2014 competition, consists of 22 layers of stacked modules with nine inception modules that include multiple convolution and pooling layers [44]. The characteristics of GoogLeNet are similar to the characteristics of the network-in-network algorithm [45], which (a) incorporates a micro-network with fully connected feature maps, instead of the activation function; (b) has a network structure that exhibits depth and width in both the vertical and horizontal dimensions; and (c) applies the rectified linear unit activation function to all convolutional layers. Both algorithms use a dropout regularization method to suppress overfitting in fully connected layers.

2.5. Learning Environment and Models for the CNNs

The learning environment for both CNNs was a computer with a Linux operating system (Ubuntu 18.04 LTS), Intel Core i5–8400K central processing unit and NVIDIA GeForce RTX 2080 Super GPU [46]. We used CUDA 11.6 and cuDNN 8.3.2 to support the GPU with deep learning. We used convolutional architecture for fast feature embedding (Caffe, ver. 0.17.3) as the learning framework for the NVIDIA deep learning GPU training system (DIGITS) 6.1.1, which enables web-based learning. Caffe is a leading deep learning framework; it has hyper-parameter tuning, rapid execution speed and supports numerous operating systems [47].

2.6. Simulation Conditions and Performance Evaluation

We divided leaf images into 10 equal sets, each of which included all 20 species. We prepared two patterns of tree species identification methods, using data augmentation and no augmentation with each pattern. The first pattern identified the 20 tree species individually (hereafter referred to as “20-species classification”); the second pattern identified the 20 tree species in two groups, broadleaf and coniferous trees (hereafter referred to as “two-group classification”). The results from a previous study [5] suggest that deep learning avoids some of the challenges associated with human classification; these two methods were implemented to confirm this suspicion. Specifically, when a human misclassifies a tree species, the misclassification typically involves a tree species with similar shape features. For example, if a tree has simple leaves with smooth margins, it is likely to be misclassified as another species that also has simple leaves and smooth margins. However, deep learning has not shown such trends [5]. Misclassification might occur because the leaf image used for deep learning is not a scanned image that consists of a single leaf (used for tree species identification by machine learning using shape features)—it is a photograph acquired in the field, in which multiple leaves are reflected. Thus, misclassification is not limited to features that arise when deep learning is used. The application of deep learning to many tree species in the future will enable species identification, but there will be limitations in that coniferous trees may be mistakenly classified as broadleaf (or broadleaf trees may be mistakenly classified as coniferous). Furthermore, rather than classifying all cases at the species level in a single instance, it may be possible to classify the trees as broadleaf or coniferous in advance, then identify the species within each group. If this method demonstrated high accuracy, it would add to the repertoire of effective methods for tree species identification models with deep learning.
Here, we devised four learning models. The datasets for all models consisted of nine sets of training data and one set of test data. Learning Model 1 (LM-1) classified 20 tree species individually without data augmentation. Learning Model 2 (LM-2) classified 20 tree species individually with data augmentation. Learning Model 3 (LM-3) classified two groups (broadleaf and coniferous trees) without data augmentation. Finally, Learning Model 4 (LM-4) classified two groups (broadleaf and coniferous trees) with data augmentation. We conducted 10 model iterations without duplication. LM-2 and LM-4 were prepared such that the 12 images enhanced by data augmentation were not divided into separate datasets (training and test); thus, the training and test datasets were independent. To evaluate the performance of the proposed models, we calculated the F-score, which was used for the classification accuracy of individual tree species (20 tree species for LM-1 and LM-2; two groups for LM-3 and LM-4) and the MCC, which was used for overall classification accuracy [48,49,50]. In a classification model that uses contingency tables, both true positive (TP) and true negative (TN) are correct classifications according to the classifier; TP is the positive example, while TN is the negative example, for each piece of training data. A false positive (FP) occurs when the outcome is incorrectly predicted to be “yes” (positive), although it is actually “no” (negative). A false negative (FN) occurs when the outcome is incorrectly predicted to be negative, although it is actually positive. The number of cases predicted to be positive that are actually positive refers to precision, while the number of cases that were actually correct and predicted to be positive is referred to as recall. Both are expressed by Equations (1) and (2), respectively.
P r e c i s i o n = T P T P + F N
R e c a l l = T P T P + F P
Both equations are indices to evaluate classification accuracy, but they involve trade-offs [49,50]. In this study, the F-score (Equation (3)), which is the harmonic mean of the precision and recall, was used as an overall indicator.
F s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
The MCC is an indicator that shows whether the classification is unbiased. It ranges from −1 to 1 [48]. The MCC was calculated using Equation (4) for LM-1 and LM-2 (because they are multi-class classifiers [51]) and Equation (5) for LM-3 and LM-4 (because they are binary classifiers). The MCC of Equation (4) can be defined by a confusion matrix C for K classes [51].
M C C = c × s k K p k × t k s 2 k K p k 2 × s 2 k K t k 2
where t k = i K C i k is the number of times class k actually occurs in the confusion matrix C of class K; p k = i K C k i is the number of times class k was predicted in the confusion matrix C of class K; c = k K C k k is the total number of samples correctly predicted; and s = i K j K C i j is the total number of samples [51].
M C C = T P × T N F P × F N T P + F P T P + F N T N + F P T N + F N
We conducted 50, 100 and 200 epochs for all learning models.

3. Results

3.1. Classification Accuracy of Tree Species Identification for Test Data When Simultaneously Identifying Broadleaf and Coniferous Trees

Table 1 shows F-scores that represent the classification accuracies of tree species for test data when broadleaf and coniferous tree species were classified simultaneously. The values in the table are the means of 10 simulation results.
Without data augmentation, the highest F-score was 0.968 for G.b. (please see abbreviations list for all species names used in this manuscript) in GoogLeNet after 200 epochs. The lowest F-score was 0.619 for T.c. in AlexNet after 50 epochs. The highest F-scores of broadleaf trees in GoogLeNet were obtained for 0.943 for C.j.Z. after 50 epochs, 0950 for C.c. after 100 epochs and 0.968 for G.b. after 200 epochs; thus, the results differed according to the number of epochs. In AlexNet, G.b. had the highest F-scores for all epochs (0.925, 0.942 and 0.941 in order of decreasing epoch). The lowest F-scores from the broadleaf tree category were L.f. for all epochs in both GoogLeNet and AlexNet (0.774, 0.846 and 0.840 in order of decreasing epochs for GoogLeNet; 0.687, 0.743 and 0.763 for AlexNet). The highest coniferous tree F-scores in GoogLeNet were P.t. for all epochs (0.890, 0.913 and 0.929 in order of decreasing epoch); the highest coniferous F-scores in AlexNet were M.g. for all epochs (0.863, 0.881 and 0.874 in order of decreasing epoch). The lowest coniferous tree F-scores in GoogLeNet were 0.751 for T.c. after 50 epochs, 0.792 for A.f. after 100 epochs and 0.812 for T.c. after 200 epochs; the lowest coniferous tree F-scores in AlexNet were T.c. for all epochs (0.619, 0.678 and 0.697 in order of decreasing epoch).
Comparison of overall F-scores by epoch revealed that minimum, maximum and mean values were higher for broadleaf trees across all epochs for both GoogLeNet and AlexNet. Similar to differences among epochs, the F-score tended to increase as epochs increased for most learning patterns, although some patterns showed slightly lower F-scores (C.c., M.t. and L.f. in GoogLeNet after 200 epochs; P.t. in AlexNet after 100 epochs; and C.j.N., M.t., S.s., G.b. and M.g. in AlexNet after 200 epochs). Among the learning algorithms used, GoogLeNet had a higher F-score for all training patterns.
With data augmentation, the highest F-score was 0.985 for Q.p. in GoogLeNet after 200 epochs, while the lowest was 0.796 for T.c. in AlexNet after 100 epochs. With regard to epoch, the highest broadleaf tree F-scores in GoogLeNet were 0.974 for C.j.Z. after 50 epochs, 0.976 for Q.p. after 100 epochs and 0.985 for Q.p. after 200 epochs; the highest broadleaf tree F-scores in AlexNet were 0.965 for C.j.Z. after 50 epochs, 0.968 for G.b. after 100 epochs and 0.972 for C.j.Z. after 200 epochs. The lowest broadleaf tree F-scores in GoogLeNet were 0.923 for L.f. after 50 epochs, 0.925 for L.g. after 100 epochs and 0.940 for L.g. after 200 epochs; the lowest broadleaf tree F-scores in AlexNet were 0.885 for L.g. after 50 epochs, 0.895 for L.g. after 100 epochs and 0.902 for L.f. after 200 epochs. The highest coniferous tree F-scores in GoogLeNet were P.t. for all epochs (0.942, 0.946 and 0.954 in order of decreasing epoch); the highest coniferous tree F-scores in AlexNet were P.t. for all epochs (0.942, 0.954 and 0.944 in order of decreasing epoch). The lowest coniferous tree F-scores in GoogLeNet were 0.842 for T.c. after 50 epochs, 0.851 for R.c. after 100 epochs and 0.871 for A.f. after 200 epochs; the lowest coniferous tree F-scores in AlexNet were T.c. for all epochs (0.812, 0.796 and 0.823 in order of decreasing epoch).
Comparison of overall F-scores by epoch revealed that minimum, maximum and mean values were higher for broadleaf trees across all epochs for both GoogLeNet and AlexNet, similar to the findings without data augmentation. Similar to differences among epochs, the F-score tended to increase as epochs increased for most of the learning patterns, but some of the patterns showed slightly lower F-scores, similar to the findings without data augmentation (A.f., C.o. and M.g. in GoogLeNet after 100 epochs; C.f., C.j.D. and T.c in AlexNet after 100 epochs; and A.p.ck, G.b., L.f., Co., C.f. and P.t. in both GoogLeNet and AlexNet after 200 epochs). Among the learning algorithms used, GoogLeNet had a higher F-score for many learning patterns; in a few cases, AlexNet had a higher score (A.f. and P.t. in AlexNet after 100 epochs).

3.2. Classification Accuracy of Tree Species Identification for Test Data When Either Broadleaf or Coniferous Trees Were Used

Table 2 shows F-scores that represent the classification accuracies of tree species for test data when classifying between groups of either broadleaf or coniferous trees. The values in the table are the means of 10 simulation results.
The highest F-score was 0.9988 for broadleaf trees with data augmentation in GoogLeNet after 100 epochs. The lowest F-score was 0.9902 for coniferous trees without data augmentation in AlexNet after 50 epochs. Overall, F-scores were very high for 20-species classification. F-scores were higher with data augmentation for all epochs. Among the various epochs, F-scores tended to increase as epochs increased, but the difference was very small. GoogLeNet after 200 epochs with data augmentation resulted in a slightly lower F-score than after 100 epochs for both broadleaf and coniferous trees. Comparison of the highest F-score for each tree species revealed almost identical results for broadleaf and coniferous trees: 0.9988 and 0.9982, respectively (both with data augmentation in GoogLeNet after 100 epochs). The number of errors was 51 of 42,000 images for broadleaf trees and 53 of 28,800 images for coniferous trees; these results indicated very high classification accuracy for group.

3.3. Classification Accuracy of Tree Species Identification According to Learning Method

Figure 3 shows the classification accuracy of tree species identification according to learning method for test data, on the basis of MCC. In the legend for Figure 3, “20 species individually” indicates LM-1 (20-species classification); “only broadleaf” indicates the MCC, which was calculated using results from Ref. [5]; “only coniferous” indicates the MCC, which was calculated using results from Ref. [6]; and “two groups” indicates LM-3 (two-group classification). Figure 3a shows GoogLeNet and Figure 3b shows AlexNet; solid lines indicate data augmentation and dashed lines indicate no data augmentation.
In Figure 3a (GoogLeNet), the highest MCC was 0.997 for “two groups, data augmentation” after 100 epochs. MCCs for “two groups, data augmentation” were high for all epochs. The lowest MCC was 0.756 for “only coniferous” after 50 epochs. Comparison of the results with and without data augmentation revealed that classification accuracy tended to improve as epochs increased in most cases without data augmentation. The two-group classifications (“two groups” and “two groups, data augmentation” in Figure 3a) showed a very high classification accuracy of 0.995 with data augmentation and 0.991 without data augmentation after 50 epochs; thus, there was almost no improvement in classification accuracy as epochs increased. With data augmentation, there was no improvement in classification accuracy as epochs increased; in some cases, there was a slight decrease in classification accuracy as epochs increased (“only coniferous, data augmentation” after 100 epochs and “two groups, data augmentation” after 200 epochs). In Figure 3b (AlexNet), the highest MCC was 0.996 for “two groups, data augmentation” after 100 epochs. MCCs for “two groups, data augmentation” were high for all epochs, similar to the GoogLeNet results. The lowest MCC was 0.684 for “only coniferous” after 50 epochs. The AlexNet results with and without data augmentation were similar to the GoogLeNet results. The classification accuracy in AlexNet was almost identical to the classification accuracy in GoogLeNet for “two groups, data augmentation”; however, the accuracy of the other groups tended to be lower overall, especially concerning MCCs without data augmentation (which were lower with the exception of “two groups”).

3.4. Comparison between Simultaneous and Individual Identification

As indicated in Section 3.3, the two-group classification had the highest accuracy, followed by only broadleaf trees, 20-species classification individually and coniferous trees only. Therefore, the accuracy of tree species identification was higher for broadleaf trees when only broadleaf trees were used; it was higher for coniferous trees when both broadleaf and coniferous trees were identified simultaneously. However, because the above MCC is calculated for the entire classification, the MCC when both broadleaf and coniferous trees were classified simultaneously may be the result of averaging the classification accuracies of broadleaf trees only and coniferous trees only. Therefore, from the cases in which broadleaf and coniferous trees were classified simultaneously, the respective results for broadleaf and coniferous trees were extracted and compared with the results from broadleaf only classification [5] and coniferous only classification [6], respectively (Table 3).
Notably, the number of misclassified tree species differed for each training model; comparisons using MCCs and F-scores were thus inappropriate. We compared the total TPs, which represent the correct classification of the positive examples, according to tree group. The values in Table 3 show the total TP value according to tree group when classified simultaneously minus the TP value of broadleaf or coniferous only, respectively. Positive (+) values indicate a higher accuracy in the classification of broadleaf and coniferous trees simultaneously, while negative (−) values indicate a higher accuracy when only broadleaf trees or only coniferous trees are classified. For broadleaf trees in GoogLeNet, all epochs with and without data augmentation showed negative values. Therefore, we presumed that the accuracy of tree species identification would be higher when learning only broadleaf trees than when learning the classification of broadleaf and coniferous trees simultaneously. Without data augmentation, the classification accuracy was lower after 100 epochs than after 50 epochs, but the overall trend was that the classification accuracy increased as epochs increased. With data augmentation, the classification accuracy tended to decrease as epochs increased. For broadleaf trees in AlexNet, the results after 50 and 100 epochs without data augmentation indicated increased classification accuracy when trees were classified simultaneously. However, classification accuracy tended to increase as epochs increased when broadleaf trees only were used for training. For coniferous trees in both GoogLeNet and AlexNet, all epochs with and without data augmentation showed positive values. Therefore, we presumed that the accuracy of tree species identification would be higher when learning the classification of broadleaf and coniferous trees simultaneously than when learning coniferous trees only. GoogLeNet showed a trend toward higher classification accuracy for coniferous trees only without data augmentation and for data augmentation with simultaneous classification as epochs increased.

3.5. Misclassification Patterns According to Tree Species

Figure 4 and Figure 5 show misclassification patterns according to tree species. The figures comprise heatmaps that were created based on the classification results obtained from the contingency table, using the seaborn module in Python. Figure values show classification according to each training model; they represent the sum of the results of 10 simulations. Based on the results in Table 3, the results with the lowest and highest MCCs are shown in Figure 4 and Figure 5, respectively. Training class represents each tree species in the training cases; test class represents where the training model classified each tree species (i.e., how the training model classified each tree species used in the training cases). The figure shows that the correct classification was made only when the training class and the test class had the same tree species. To clarify the misclassification, broadleaf and coniferous trees are indicated along with three groups of broadleaf trees based on leaf shape, which are divided and clearly indicated using solid lines.
In Figure 4, the highest TP was G.b. and the lowest TP was T.c. Overall, there were few misclassifications of broadleaf trees as coniferous trees or coniferous trees as broadleaf trees. Both broadleaf and coniferous trees were frequently misclassified within each category; broadleaf trees were misclassified within three categories. The four species of broadleaf trees that were not misidentified as coniferous trees were M.t., C.j.N., Q.g. and A.p.ck. There were no cases in which coniferous trees were misidentified as broadleaf trees among any of the tree species. Except for T.c., seven coniferous trees had no misclassification of broadleaf trees as a category of simple leaves with toothed margins. Comparison of the four categories (three broadleaf categories and one coniferous) revealed that the categories with the highest total number of misclassifications for each category were the misclassification of simple leaves with smooth margins as simple leaves with toothed margins (99 total misclassifications), simple leaves with toothed margins as simple leaves with smooth margins (130), lobed leaves as coniferous leaves (119) and coniferous leaves as lobed leaves (568).
In Figure 5, the highest TP was S.s. and the lowest TP was A.f. Similar to Figure 4, there were few cases of misclassification of broadleaf trees as coniferous trees or coniferous trees as broadleaf trees. Both broadleaf and coniferous trees were frequently misclassified within each category; broadleaf trees were misclassified within three categories. The five species for which broadleaf trees were not misidentified as coniferous trees were M.t., C.j.N., Q.g., Q.p. and A.p.ck. The three species for which coniferous trees were not misidentified as broadleaf trees were C.o., M.g. and T.c. There were no misclassifications for simple leaves with toothed margins and lobed leaves on A.f. and P.t. or for simple leaves with smooth leaves and lobed leaves on C.f. Comparison of the four categories of broadleaf and three categories of coniferous trees revealed that the categories with the highest total number of misclassifications for each category were the misclassification of simple leaves with smooth margins as simple leaves with toothed margins (186), simple leaves with toothed margins as simple leaves with toothed margins (168), lobed leaves as simple leaves with smooth margins (245) and coniferous leaves as lobed leaves (2339).

4. Discussion

In this study, when 12 broadleaf and eight coniferous trees were classified simultaneously, the highest MCC was 0.8914 without data augmentation and 0.9447 with data augmentation. While the model with the highest MCC for only 12 broadleaf trees had a classification accuracy of 0.9309 without data augmentation and 0.9670 with data augmentation [5], only eight coniferous trees had a classification accuracy of 0.8465 without data augmentation and 0.8695 with data augmentation [6]. Considering the results in Table 3, the classification accuracy was higher for broadleaf trees when the model was trained using broadleaf only, whereas the classification accuracy of coniferous trees was higher when the model was trained using both tree categories simultaneously than when it was trained with coniferous trees only. These findings should be verified in additional studies because of the difference in the number of broadleaf and coniferous trees; moreover, the leaf images were acquired from only a few trees of each species.
With respect to epochs, many models showed a trend toward improved classification accuracy as epochs increased. Especially without data augmentation, there was substantial improvement in classification accuracy as epochs increased. While high classification accuracy was achieved after 50 epochs (the number of epochs used as the fewest number of training sessions), there was minimal improvement in classification accuracy as epochs increased thereafter; there were a few cases in which classification accuracy slightly decreased. In most types of machine learning, including deep learning, the use of excessive learning iterations (i.e., epochs in deep learning) can lead to low classification accuracy of test data because the training data are overfitted to the training cases; this is known as “overfitting” [52]. In the present study, overfitting presumably occurred in the two-group classification as epochs increased because these models achieved sufficient classification accuracy with a small number of epochs.
With respect to learning algorithms, GoogLeNet tended to have higher classification accuracy than AlexNet for most of the models. Because of the similar results in tree species identification in 2015 PlantCLEF and Ghazi et al. [17] mentioned above, GoogLeNet is regarded as one of the most effective deep learning algorithms with high classification accuracy. A possible reason for GoogLeNet to exhibit higher classification accuracy is that GoogLeNet has more layers than AlexNet. Deep learning tends to improve image recognition performance as the number of layers increases. However, several challenges have been identified in the application of deep learning. Constraints on input data (i.e., image type and image size) vary among algorithms; moreover, algorithms with many layers require extensive computation as the number of layers increases, which leads to long training times or the need for high performance hardware. Therefore, it is preferable to consider the optimal type of learning model for different problem sets, rather than simply using a model with many layers.
With respect to data augmentation, the classification accuracy tended to improve with data augmentation when many objects were identified. There was also sufficient classification accuracy without data augmentation when identifying both broadleaf and coniferous trees (two-group classification). This finding is presumably because a large amount of training data was used for two-group classification: 3600 images for broadleaf trees and 2400 images for coniferous trees. If high classification accuracy was achieved without data augmentation, data augmentation led to a slight decrease in classification accuracy. In general, deep learning requires the preparation of a large amount of training data for analysis, but there is no specific threshold for analysis target data. Based on our results, determination of the need for data augmentation should be based on the classification target and the amount of data. If classification accuracy is high despite a small amount of training data, data augmentation is unnecessary; in such cases, an unnecessarily large number of epochs can decrease classification accuracy because of overfitting.
With respect to misclassification patterns, broadleaf trees were often misclassified as the incorrect broadleaf species and coniferous trees were often misclassified as the incorrect coniferous species. However, because identification is not 100% correct, there were a few cases where broadleaf trees were misclassified as coniferous trees or coniferous trees were misclassified as broadleaf trees; such misclassification is unlikely to occur with human identification. Although detailed verification was not conducted in this study because it would been extremely complicated, image recognition using deep learning can output classification results as rankings, as in the mean reciprocal rank method described earlier, and in the DIGITS method used in this study, where the classification results are displayed as probabilities with the top five rankings. Therefore, when constructing a practical tree species identification system, it is preferable to use a method that can present several candidates instead of a single tree species. In addition, misidentifications (e.g., confusion between broadleaf and coniferous trees) can be prevented if candidate tree species are expressed as probabilities or if images of tree species and textual information about tree species are presented simultaneously.
Based on the results of this study, we propose the following learning method for constructing a tree species identification system. First, distinguish between two groups: broadleaf and coniferous trees. Then, perform classification of broadleaf trees with a learning model for broadleaf trees only; this will provide higher classification accuracy. For coniferous trees, use a classification model for broadleaf and coniferous trees simultaneously; this will provide higher classification accuracy. Notably, we found distinct trends in species identification between broadleaf and coniferous trees, but the reasons for these differences are unclear. The number of tree species used in this study was small; we expect that the trend would have differed if more tree species had been used. To examine the accuracy of tree species identification using deep learning, we investigated deep learning algorithms, learning patterns, epochs and the amounts of training and test data with and without data augmentation. To improve the accuracy of tree species classification, there is a need to improve the classification accuracy of coniferous trees in particular. Considering that the species classification accuracy is lower for coniferous trees than for broadleaf trees (although the number of coniferous species is smaller than the number of broadleaf tree species), learning conditions should be considered along with other conditions (e.g., learning using leaf images, as well as images of whole trees and stems). Furthermore, we are exploring the construction of a system that can perform tree species identification by implementing the results of machine and deep learning identification that have been conducted on mobile terminals. To achieve tree species identification using mobile terminals, machine and deep learning must be implemented on the mobile terminals. We are conducting tree species identification using deep learning for mobile terminals such as MobileNetV3 [53] and GhostNet [54], which are typical deep learning algorithms. Alternatively, rather than direct implementation on a mobile device, CoreML for iOS [37] and ML Kit for Android [55] have been proposed for training on non-mobile devices; these softwares allow the results to subsequently be applied to a mobile device. Empirical studies using these methods will be important in the future.

5. Conclusions and Future Work

We applied a deep learning technique to identify tree species based on outdoor photographs of leaves from 12 broadleaf and eight conifer species. We evaluated the accuracy of our results by classifying broadleaf and conifer trees simultaneously. Outdoor tree leaf photographs typically include multiple leaves, rather than single leaves. Although single-leaf images or those adjusted for brightness or saturation are associated with high tree species classification accuracy [30,31,32,33,34], our results show that multiple-leaf images acquired in the field could be used to classify tree species with high accuracy. Variation in classification accuracy and misclassification was observed for both broadleaf and conifer trees. Further study is required to verify the results of this study using a larger number of tree species. In this study, we did not examine the effects of the conditions under which photographs were taken. It is possible that different backgrounds for the same tree species in various locations may have affected classification accuracy. For example, a previous study found that leaf segmentation or overlapping strongly influenced the classification of broadleaf tree images, particularly in images with complex backgrounds, necessitating labeling for segmentation [32]. We conclude that labeling should also be considered for coniferous trees, although it would increase the labor required for image processing.

Author Contributions

Conceptualization, Y.M.; methodology, Y.M.; software, Y.M.; validation, Y.M., Y.K. and S.N.; formal analysis, Y.M., Y.K. and S.N.; investigation, Y.K. and S.N.; resources, Y.K. and S.N.; data curation, Y.M., Y.K. and S.N.; writing—original draft preparation, Y.M., Y.K. and S.N.; writing—review and editing, Y.M.; visualization, Y.M.; supervision, Y.M.; project administration, Y.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to the members of the Laboratory of Watershed Information Science and the Laboratory of Erosion Control, Kyoto Prefectural University, for their cooperation in conducting this study.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNconvolutional neural network
MCCMatthews correlation coefficient
TPtrue positive
TNtrue negative
FPfalse positive
FNfalse negative
C.c.Cinnamomun camphora (Linn.) Sieb.
C.j.NCinnamomum japonicum Sieb. ex Nakai
M.t.Maclura tricuspidata (syn. Cudrania tricuspidata)
S.s. Sapium sebiferum (Linn.) Roxb.
C.j.ZCercidiphyllum japonicum Sieb. et Zucc.
L.g. Lithocarpus glaber Thunb.
Q.g.Quercus glauca Thunb.
Q.p.Quercus philyraeoides A. Gray
A.p.Acer palmatum Thunb.
A.p.ckAcer platanoides cv. crimson king
G.b.Ginkgo biloba Linn.
L.f.Liquidambar formosana Hance.
A.f.Abies firma
C.f.Calocedrus formosana
C.j.DCryptomeria japonica D. Don
C.o.Chamaecyparis obtusa
M.g.Metasequoia glyptostroboides
P.t.Pinus thunbergii
T.c.Taxus cuspidata
T.s.Tsuga sieboldii

References

  1. Minowa, Y.; Akiba, T.; Nakanishi, Y. Classification of a leaf image using a self-organizing map and tree based model. J. For. Plan. 2011, 17, 31–42. [Google Scholar] [CrossRef]
  2. Minowa, Y.; Takami, T.; Suguri, M.; Ashida, M.; Yoshimura, Y. Identification of tree species using a machine learning algorithm based on leaf shape and venation pattern. Jpn. J. Plann 2019, 53, 1–13. [Google Scholar] [CrossRef]
  3. Minowa, Y.; Asao, K. Tree species identification based on venation pattern of leaf images photographed with a mobile device in the outdoors. Jpn. J. Plann 2020, 53, 43–52. [Google Scholar] [CrossRef]
  4. Minowa, Y.; Nagasaki, Y. Convolutional neural network applied to tree species identification based on leaf images. J. For. Plan. 2020, 26, 1–11. [Google Scholar] [CrossRef]
  5. Minowa, Y.; Kubota, Y. Identification of broad-leaf trees using deep learning based on field photographs of multiple leaves. J. For. Res. 2022, 1–9. [Google Scholar] [CrossRef]
  6. Minowa, Y.; Nakatsukasa, S. Identification of coniferous tree species using deep learning. Jpn J. Plann 2021, 13, 162. [Google Scholar]
  7. Wang, Z.; Chi, Z.; Feng, D.; Wang, Q. Leaf image retrieval with shape features. Lect. Notes Comput. Sci. 2000, 1929, 477–487. [Google Scholar]
  8. Nam, Y.; Hwang, E. A shape-based retrieval scheme for leaf image. Lec. Notes Comput. Sci. 2005, 3767, 876–887. [Google Scholar]
  9. Shen, Y.; Zhou, C.; Lin, K. Leaf Image Retrieval Using a Shape Based Method. In Proceedings of the IFIP International Conference on Artificial Intelligence Applications and Innovations, Boston, MA, USA, 7–9 September 2005; pp. 711–719. [Google Scholar]
  10. Neto, J.C.; Meyer, G.E.; Jones, D.D.; Samal, A.K. Plant species identification using elliptic Fourier leaf shape analysis. Comput. Electron. Agric. 2006, 50, 121–134. [Google Scholar] [CrossRef]
  11. Du, J.X.; Wang, X.; Zhang, G.J. Leaf shape based plant species recognition. Appl. Math. Comput. 2007, 185, 883–893. [Google Scholar] [CrossRef]
  12. Beghin, T.; Cope, J.S.; Remagnino, P.; Barman, S. Shape and texture based plant leaf classification. Lect. Notes Comput. Sci. 2010, 6475, 345–353. [Google Scholar]
  13. Aptoula, E.; Yanikoglu, B. Morphological features for leaf based plant recognition. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 1496–1499. [Google Scholar]
  14. Li, Y.; Chi, Z.; Feng, D.D. Leaf vein extraction using independent component analysis. In Proceedings of the 2006 IEEE International Conference on Systems, Man and Cybernetics, Taipei, Taiwan, 8–11 October 2006; pp. 3890–3894. [Google Scholar]
  15. Cope, J.S.; Remagnino, P.; Barman, S.; Wilkin, P. The extraction of venation from leaf images by evolved vein classifiers and ant colony algorithms. In Proceedings of the International Conference on Advanced Concepts for Intelligent Vision Systems, Sydney, Australia, 13–16 December 2010; Volume 6474, pp. 135–144. [Google Scholar]
  16. Wilf, P.; Zhang, Z.; Chikkerur, S.; Little, S.A.; Wing, S.L.; Serre, T. Computer vision cracks the leaf code. Proc. Natl. Acad. Sci. USA 2016, 113, 3305–3310. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Ghazi, M.M.; Yanikoglu, B.; Aptoula, E. Plant identification using deep neural networks via optimization of transfer learning parameters. Neurocomputing 2017, 235, 228–235. [Google Scholar] [CrossRef]
  18. Cope, J.S.; Corney, D.P.A.; Clark, J.Y.; Remagnino, P.; Wilkin, P. Plant species identification using digital morphometrics: A review. Expert. Syst. Appl. 2012, 39, 7562–7573. [Google Scholar] [CrossRef]
  19. Yamashita, T. Irasuto de Manabu Deep Learning [An Illustrated Guide to Deep Learning]; Kodansha: Tokyo, Japan, 2016. [Google Scholar]
  20. Deng, Y. Deep learning on mobile devices—A review. Mob. Multimed. /Image Process. Secur. Appl. 2019, 10993, 52–66. [Google Scholar]
  21. Lin, T.Y.; Maire, M.; Belongie, S.; Bourdev, L.; Girshick, R.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L.; Dollár, P. Microsoft COCO: Common Objects in Context; Springer: Tokyo, Japan, 2014; Volume 8693, pp. 740–755. [Google Scholar]
  22. Goëau, H.; Bonnet, P.; Joly, A. LifeCLEF plant identification task 2015. CLEF (Working Notes). In Proceedings of the Working Notes for CLEF 2014 Conference, Sheffield, UK, 15–18 September 2014. [Google Scholar]
  23. Zhou, B.; Lapedriza, A.; Khosla, A.; Oliva, A.; Torralba, A. Places: A 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1–14. [Google Scholar] [CrossRef] [Green Version]
  24. Torrey, L.; Shavlik, J. Transfer Learning. 2009. Available online: https://ftp.cs.wisc.edu/machine-learning/shavlik-group/torrey.handbook09.pdf (accessed on 25 August 2021).
  25. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep neural networks based recognition of plant diseases by leaf image classification. Comput. Intell. Neurosci. 2016, 2016, 1–11. [Google Scholar] [CrossRef] [Green Version]
  26. Boulent, J.; Foucher, S.; Théau, J.; St-Charles, P. Convolutional neural networks for the automatic identification of plant diseases. Front. Plant Sci. 2019, 40, 941. [Google Scholar] [CrossRef] [Green Version]
  27. Yun, S.; Donghui, M.; Jie, L.; Jie, L. ACTL: Asymmetric convolutional transfer learning for tree species identification based on deep neural network. IEEE Access 2021, 9, 13643–13654. [Google Scholar]
  28. Li, H.; Hu, B.; Li, Q.; Jing, L. CNN-based individual tree species classification using high-resolution satellite imagery and airborne LiDAR data. Forests 2021, 12, 1697. [Google Scholar] [CrossRef]
  29. Nezami, S.; Khoramshahi, E.; Nevalainen, O.; Pölönen, I.; Honkavaara, E. Tree species classification of drone hyperspectral and RGB imagery with deep learning convolutional neural networks. Remote Sens. 2020, 12, 1070. [Google Scholar] [CrossRef] [Green Version]
  30. Bisen, D. Deep convolutional neural network based plant species recognition through features of leaf. Multimed. Tools Appl. 2020, 80, 6443–6456. [Google Scholar] [CrossRef]
  31. Gyires-Tóth, B.P.; Osváth, M.; Papp, D.; Szűcs, G. Deep learning for plant classification and content-based image retrieval. Cybern. Inf. Technol. 2019, 19, 88–100. [Google Scholar] [CrossRef] [Green Version]
  32. Yang, K.; Zhong, W.; Li, L. Leaf segmentation and classification with a complicated background using deep learning. Agronomy 2020, 10, 1721. [Google Scholar] [CrossRef]
  33. Hamrouni, L.; Kherfi, M.; Lamine, A.; Oussama, B.; Abdellah. Plant leaves recognition based on a hierarchical one-class learning scheme with convolutional auto-encoder and siamese neural network. Symmetry 2021, 13, 1705. [Google Scholar] [CrossRef]
  34. Quoc Bao, T.; Tan Kiet, N.T.; Quoc Dinh, T.; Hiep, H.X. Plant species identification from leaf patterns using histogram of oriented gradients feature space and convolution neural networks. J. Inf. Telecommun. 2020, 4, 140–150. [Google Scholar] [CrossRef] [Green Version]
  35. ImageCLEF2012. 2012. Available online: https://www.imageclef.org/2012/plant/ (accessed on 8 June 2022).
  36. Swedish Leaf Dataset. 2016. Available online: https://www.cvl.isy.liu.se/en/research/datasets/swedish-leaf/ (accessed on 8 June 2022).
  37. CoreML. 2017. Available online: https://developer.apple.com/documentation/coreml/ (accessed on 19 August 2021).
  38. Tsutsumi, S.; (iOS Engineer, Freelance, Tokyo, Japan). Personal Private communication, 2020.
  39. Nikon COOLPIX A900. 2016. Available online: https://www.nikon-image.com/products/compact/lineup/a900/ (accessed on 3 August 2021).
  40. Samsung Galaxy S9. 2018. Available online: https://www.samsung.com/global/galaxy/galaxy-s9/ (accessed on 5 January 2022).
  41. NIH. ImageJ. 2004. Available online: https://imagej.nih.gov/ij/ (accessed on 19 February 2021).
  42. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
  43. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Processing Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  44. Szegedy, C.; Liu, W.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7 June 2015; pp. 1–12. [Google Scholar]
  45. Liu, M.; Chen, Q.; Yan, S. Network in network. arXiv 2014, arXiv:1312.4400. [Google Scholar]
  46. NVIDIA. 2016. Available online: https://developer.nvidia.com/cuda-toolkit/ (accessed on 19 February 2021).
  47. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional architecture for fast feature embedding. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 675–678. [Google Scholar]
  48. Matthews, B.W. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Et Biophys. Acta (BBA)—Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  49. Motoda, H.; Tsumoto, S.; Yamaguchi, T.; Numao, M. Fundamentals of Data Mining; Ohmsha: Tokyo, Japan, 2006. [Google Scholar]
  50. Witten, I.H.; Frank, E.; Hall, M.A. Data Mining, Practical Machine Learning Tools and Techniques, 3rd ed.; Morgan Kaufmann Publishers: Burlington, NJ, USA, 2011; pp. 163–180. [Google Scholar]
  51. Scikit-Learn. 2007. Available online: https://scikit-learn.org/stable/modules/model_evaluation.html#matthews-corrcoef (accessed on 5 February 2022).
  52. Raschka, S.; Mirjalili, V. Python Machine Learning Programming; Impress: Tokyo, Japan, 2018. [Google Scholar]
  53. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for MobileNetV3. arXiv 2019, arXiv:1905.02244. [Google Scholar]
  54. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More features from cheap operations. arXiv 2020, arXiv:1911.11907. [Google Scholar]
  55. ML Kit. 2018. Available online: https://developers.google.com/ml-kit/ (accessed on 19 August 2021).
Figure 1. Focal species and classification based on broadleaf and coniferous trees.
Figure 1. Focal species and classification based on broadleaf and coniferous trees.
Forests 13 00943 g001
Figure 2. Methods for creating leaf images from photographs.
Figure 2. Methods for creating leaf images from photographs.
Forests 13 00943 g002
Figure 3. Classification accuracy of tree species identification for test data, based on MCC. (a) GoogLeNet; (b) AlexNet.
Figure 3. Classification accuracy of tree species identification for test data, based on MCC. (a) GoogLeNet; (b) AlexNet.
Forests 13 00943 g003
Figure 4. Misclassification patterns according to tree species in AlexNet after 50 epochs without data augmentation. 1 Simple leaves with smooth margins, 2 Simple leaves with toothed margins, 3 Lobed leaves.
Figure 4. Misclassification patterns according to tree species in AlexNet after 50 epochs without data augmentation. 1 Simple leaves with smooth margins, 2 Simple leaves with toothed margins, 3 Lobed leaves.
Forests 13 00943 g004
Figure 5. Misclassification patterns according to tree species in GoogLeNet after 200 epochs with data augmentation. 1 Simple leaves with smooth margins, 2 Simple leaves with toothed margins, 3 Lobed leaves.
Figure 5. Misclassification patterns according to tree species in GoogLeNet after 200 epochs with data augmentation. 1 Simple leaves with smooth margins, 2 Simple leaves with toothed margins, 3 Lobed leaves.
Forests 13 00943 g005
Table 1. Classification accuracy of tree species identification for test data when simultaneously classifying broadleaf and coniferous trees.
Table 1. Classification accuracy of tree species identification for test data when simultaneously classifying broadleaf and coniferous trees.
CategoriesTree SpeciesNo Data AugmentationData Augmentation
GoogLeNetAlexNetGoogLeNetAlexNet
50 1100 1200 150 1100 1200 150 1100 1200 1501100 1200 1
Broadleaf treesSmooth margins 2C.c.0.9120.9500.9480.8990.9140.9230.9660.9720.9750.9490.9560.960
C.j.N0.8810.8980.9060.8020.8460.8390.9590.9620.9690.9160.9250.935
M.t.0.8460.9170.9140.8100.8590.8500.9530.9640.9740.9420.9440.956
S.s.0.8700.9160.9290.7770.8450.8440.9610.9630.9770.9310.9430.947
Toothed margins 3C.j.Z0.9430.9460.9620.9130.9390.9400.9740.9750.9810.9650.9660.972
L.g.0.7880.8470.8720.7400.7610.7660.9240.9250.9400.8850.8950.904
Q.g.0.8630.9040.9080.8080.8330.8340.9520.9620.9690.9280.9450.947
Q.p.0.8410.9260.9280.7970.8610.8850.9690.9760.9850.9440.9540.958
Lobed 4A.p.0.7930.8590.8680.7350.7640.7770.9410.9440.9540.9000.9080.910
A.p.ck0.8260.9080.9180.8210.8490.8560.9650.9660.9770.9400.9590.953
G.b.0.9290.9490.9680.9250.9420.9410.9640.9690.9720.9560.9680.963
L.f.0.7740.8460.8400.6870.7430.7630.9230.9270.9450.8990.9050.902
Min.0.7740.8460.8400.6870.7430.7630.9230.9250.9400.8850.8950.902
Max.0.9430.9500.9680.9250.9420.9410.9740.9760.9850.9650.9680.972
Ave.0.8550.9060.9140.8090.8460.8520.9540.9590.9680.9300.9390.942
Coniferous treesA.f.0.7560.7920.8180.6730.7290.7310.8750.8570.8710.8490.8610.869
C.f.0.8050.8120.8470.7650.7700.7760.8570.8670.8730.8460.8430.833
C.j.D0.7680.8290.8610.7450.7500.7580.9200.9250.9300.9090.9030.904
C.o.0.8260.8620.8830.6960.7620.7760.9290.9230.9350.8800.8970.886
M.g.0.8720.9120.9170.8630.8810.8740.9290.9180.9340.9070.9110.914
P.t.0.8900.9130.9290.8370.8260.8450.9420.9460.9540.9420.9540.944
T.c.0.7510.8010.8120.6190.6780.6970.8420.8510.8830.8120.7960.823
T.s.0.8340.8880.8990.7390.7840.8130.9310.9420.9520.9000.9030.907
Min.0.7510.7920.8120.6190.6780.6970.8420.8510.8710.8120.7960.823
Max.0.8900.9130.9290.8630.8810.8740.9420.9460.9540.9420.9540.944
Ave.0.8130.8510.8710.7420.7720.7840.9030.9040.9170.8800.8840.885
Min. for overall0.7510.7920.8120.6190.6780.6970.8420.8510.8710.8120.7960.823
Max. for overall0.9430.9500.9680.9250.9420.9410.9740.9760.9850.9650.9680.972
Ave. for overall0.8380.8840.8960.7820.8170.8240.9340.9370.9470.9100.9170.919
1 Epochs, 2 Simple leaves with smooth margins, 3 Simple leaves with toothed margins, 4 Lobed leaves.
Table 2. F-scores as the classification accuracy of tree species for test data when classifying between groups of either broadleaf or coniferous trees.
Table 2. F-scores as the classification accuracy of tree species for test data when classifying between groups of either broadleaf or coniferous trees.
GroupsData AugmentationGoogLeNetAlexNet
50 1100 1200 150 1100 1200 1
Broadleaf -0.99650.99670.99670.99350.99530.9960
20.99810.99880.99820.99780.99830.9981
Coniferous -0.99480.99500.99500.99020.99290.9940
20.99720.99820.99740.99670.99740.9972
1 Epochs, 2 Open circles indicate with data augmentation.
Table 3. Comparison between simultaneous identification of broadleaf and coniferous trees and individual identification of broadleaf and coniferous trees.
Table 3. Comparison between simultaneous identification of broadleaf and coniferous trees and individual identification of broadleaf and coniferous trees.
PatternData AugmentationGoogLeNetAlexNet
50 1100 1200 150 1100 1200 1
SB 2-−31−24−843520−6
4−399−275−76−470−150−212
SC 3-653916491430
4772945900462536294
1 Epochs, 2 Total value of TP according to tree group when classified simultaneously minus the TP value of broadleaf trees only, 3 Total value of TP according to tree group when classified simultaneously minus the TP value of coniferous trees only, 4 Open circles indicate with data augmentation.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Minowa, Y.; Kubota, Y.; Nakatsukasa, S. Verification of a Deep Learning-Based Tree Species Identification Model Using Images of Broadleaf and Coniferous Tree Leaves. Forests 2022, 13, 943. https://0-doi-org.brum.beds.ac.uk/10.3390/f13060943

AMA Style

Minowa Y, Kubota Y, Nakatsukasa S. Verification of a Deep Learning-Based Tree Species Identification Model Using Images of Broadleaf and Coniferous Tree Leaves. Forests. 2022; 13(6):943. https://0-doi-org.brum.beds.ac.uk/10.3390/f13060943

Chicago/Turabian Style

Minowa, Yasushi, Yuhsuke Kubota, and Shun Nakatsukasa. 2022. "Verification of a Deep Learning-Based Tree Species Identification Model Using Images of Broadleaf and Coniferous Tree Leaves" Forests 13, no. 6: 943. https://0-doi-org.brum.beds.ac.uk/10.3390/f13060943

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop