Next Article in Journal
Range-Doppler Based CFAR Ship Detection with Automatic Training Data Selection
Next Article in Special Issue
KLUM: An Urban VNIR and SWIR Spectral Library Consisting of Building Materials
Previous Article in Journal
Method of Delimiting the Spatial Structure of Villages for the Purposes of Land Consolidation and Exchange
Previous Article in Special Issue
Extraction of Urban Objects in Cloud Shadows on the basis of Fusion of Airborne LiDAR and Hyperspectral Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of Hyperspectral Techniques for Urban Tree Diversity Classification

1
CNRS, Université Rennes 2, Unité Mixte de Recherche 6554 LETG, Place du Recteur Henri le Moal, 35043 Rennes CEDEX, France
2
CNRS, Université de Rennes 2, Unité Mixte de Recherche 6554 LETG, Place Recteur Henri le Moal, CS 24307-35043 Rennes CEDEX, France
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(11), 1269; https://0-doi-org.brum.beds.ac.uk/10.3390/rs11111269
Submission received: 28 March 2019 / Revised: 5 May 2019 / Accepted: 23 May 2019 / Published: 28 May 2019
(This article belongs to the Special Issue Hyperspectral Imagery for Urban Environment)

Abstract

:
This research aims to assess the capabilities of Very High Spatial Resolution (VHSR) hyperspectral satellite data in order to discriminate urban tree diversity. Four dimension reduction methods and two classifiers are tested, using two learning methods and applied with four in situ sample datasets. An airborne HySpex image (408 bands/2 m) was acquired in July 2015 from which prototypal spaceborne hyperspectral images (named HYPXIM) at 4 m and 8 m and a multispectral Sentinel2 image at 10 m have been simulated for the purpose of this study. A comparison is made using these methods and datasets. The influence of dimension reduction methods is assessed on hyperspectral (HySpex and HYPXIM) and Sentinel2 datasets. The influence of conventional classifiers (Support Vector Machine –SVM– and Random Forest –RF–) and learning methods is evaluated on all image datasets (reduced and non-reduced hyperspectral and Sentinel2 datasets). Results show that HYPXIM 4 m and HySpex 2 m reduced by Minimum Noise Fraction (MNF) provide the greatest classification of 14 species using the SVM with an overall accuracy of 78.4% (±1.5) and a kappa index of agreement of 0.7. More generally, the learning methods have a stronger influence than classifiers, or even than dimensional reduction methods, on urban tree diversity classification. Prototypal HYPXIM images appear to present a great compromise (192 spectral bands/4 m resolution) for urban vegetation applications compared to HySpex or Sentinel2 images.

Graphical Abstract

1. Introduction

Urbanization reflects important environmental transformations, such as a change in the energy balance [1], ecosystem loss or fragmentation, air, soil and water contamination, or even a loss of farming land, as well as an increase in the need for water and a reduction of biodiversity [2,3,4,5]. In addition, urban areas are extremely sensitive to global change as they have higher temperatures and atmospheric concentrations of pollutants or fine particles than rural areas [6].
In this context, urban vegetation has several interests: it provides many services such as reducing the urban heat island by providing shade, the evapotranspiration and the photosynthesis of the trees [7,8,9] and improving air quality or even urban biodiverse habitats [8]. Urban vegetation also presents drawbacks such as the exacerbation of allergies due to pollen, and the provision of habitats for certain bird species considered harmful by the population. However, these services and disservices depend on the plant species, their location and structure [6,10,11,12]. For instance, their capacity to mitigate the urban heat island depends on the structure and composition of the species on the one hand, and on their respective water needs on the other hand [13,14]. Regarding these advantages and drawbacks, mapping urban tree diversity is therefore an important issue; herein diversity is defined as the greatest number of species or families that can be distinguished.
Remote sensing applied in urban areas offers the opportunity to overcome a lack of reliable and reproducible vegetation information in private and public lands to more accurately assess ecosystem functions [15,16,17,18]. Some authors show the value of VHSR multispectral data such as Pléiades or Ikonos to distinguish tree species [19,20,21]. For instance, Pu et al. [15] obtained an overall accuracy (OA) of 67.22% with WorldView 2 (eight bands), although the producer’s accuracy was not good enough due to this low spectral resolution. Multispectral data such as Sentinel (10–20 m spatial resolution) show also great potential for tree species classification [22,23]. Immitzer et al. [22] obtained an OA of 65% for seven tree species without exploring the full potential of Sentinel2 high repetitiveness. Hyperspectral data also tends to develop for the identification of urban vegetation. While this technology is very spectrally accurate (more than 100 contiguous and narrow bands of about 10 nm each), the spatial resolution of hyperspectral satellite data remains insufficient. Conversely, airborne hyperspectral sensors have demonstrated their capability for tree classification [16,17,18,19,20,21,22,23,24] (Table 1). Performance is stronger when merged with LiDAR data [17,18,24,25] using three-dimensional information. For example, Alonzo et al. [17] ranked 29 species with an OA of 79.2% using hyperspectral data alone, and 83.4% with additional Lidar data.
Although the large number of bands of hyperpsectral imagery can contribute to the better discrimination and classification of spectral signatures [28], the induced high dimensionality is one of the main limitations of hyperspectral images. Too numerous spectral bands may provide more confusion than separability, making statistical methods less accurate than with a limited number of spectral bands—this is known as the Hugues phenomenon [29]. This has contributed to the development of dimension reduction methods aimed at selecting the most relevant information. Used to pre-process hyperspectral imageries [30,31], dimension reduction (DR) methods project spectral information into a smaller subspace in order to minimize the amount of data while maximizing the statistical separability of spectral values [32]. Two main DR methods can be distinguished:
  • Feature extraction is defined as a set of methods for extracting information from an image [33], such as transformative methods to derive a characteristic [34]. Principal component analysis (PCA) is often used as a DR method because it preserves the greatest spectral variability of original images [35]. However, it does not take account of the signal-to-noise ratio (SNR), unlike the minimal noise fraction (MNF) [36].
  • Feature selection methods consist of considering one or more subsets of initial features according to their relevance to the considered question. These methods preserve the interpretability of the original data [37]. In the absence of prior knowledge, a selection is made based on the most informative, least correlated elements in order to preserve the original information as much as possible [38,39]. Vegetation indices are regularly used to provide information on the biogeochemical properties of plants in order to distinguish them. Moreover, these indices allow for better discrimination than when using spectral values from spectral bands alone, for instance in the case of a vegetation class whose composition is too similar [37].
Other factors influence the tree species classification. Beker et al. [40] showed that pixel size strongly influences the accuracy of the classification (for wetland mapping). Further, Dalponte et al. [28] found that the influence of spectral resolution is strongly related to the choice of a classifier. The support vector machine (SVM) and random forest (RF) classifiers are often used for tree species classification with hyperspectral data and are described as the best classifiers for pure pixels [41,42,43,44].
The high-dimensional nature of hyperspectral data introduces significant limitations for supervised classification, including the availability of learning samples [45]. Regularly, the number of samples per class is much smaller than the dimensionality of the problem, resulting in weak classification results [46]. One approach is to recognize that individual training samples may affect the classification differently and therefore use the most informative samples [44]. When a limited number of training samples is available, the stratified K-folds cross-validation is often used to split the dataset into training and validation samples, taking into account the proportion of individuals from each class. It is assimilated as one of the most rigorous methods to prevent overfitting and to estimate properly the error rate [47]. But K-folds results depend of the training set and the partition into folds [47]. Alonzo et al. [48] and Clark et al. [38] use stratified 10-fold cross-validation for an unbalanced and limited number of training samples, showing less biased results than with 2-folds while remaining less costly with respect to computation when the number of folds is equal to the number of samples.
This state-of-the-art technology shows that the accuracy of tree species classification using VHRS hyperspectral images can be influenced by DR methods, classifiers, learning methods, defined hereafter as hyperspectral techniques. In this context, evaluating the capabilities of VHSR hyperspectral sensors in order to classify the diversity of tree species in urban area is of prime importance. The aim of this study is to compare and assess which hyperspectral techniques, or combinations, show the greatest performance. Moreover, as spatial and spectral resolutions may influence classification accuracy, another objective consists of assessing which resolution is the most appropriate for urban tree diversity classification. Indeed, most of the listed literature (Table 1) used airborne hyperspectral data with a spatial resolution below 4 m width, which might not be achievable for spaceborne hyperspectral sensors. This study would thus contribute to the design of a future spaceborne hyperspectral sensor (HYPXIM) by assessing the optimal spatial resolutions for urban applications. While limited to tree diversity classification in this study, it has already been assessed for the detection of photovoltaic panels for example [49]. This paper focuses on commonly-used hyperspectral techniques as their respective influence is rarely assessed. While focusing mainly on hyperspectral data and techniques, the comparison with results obtained with Sentinel2 data may emphasize the added-value of the former.
The first part of this article presents the case study of Toulouse, on which this study is focused, the hyperspectral and ground-truth dataset, and all the methods used. The second part describes the results of experiments made to assess the respective and combined influence of hyperspectral techniques. The third part discusses possible perspectives and improvements for tree detection in urban environments, as well as the potential of spaceborne VHRS hyperspectral imaging for the study of urban areas in general.

2. Materials and Methods

2.1. Study Area

The studied site is the city center of Toulouse (43°36’16’’N, 1°26’39’’E), Southwest France (Figure 1). The urban area is strongly artificialized and the city center is characterized by vegetation present both in public areas (along road and river networks such as the nearby Canal du midi and the Grand-Rond roundabout) and in private areas (inner courtyards, private gardens) nestled between relatively dense buildings.

2.2. Remote Sensing Data

As shown in the introduction, spaceborne hyperspectral imagery can be very convenient for urban applications. The French National Space Center (CNES) is considering this option by assessing the potential of such a sensor: the HYPXIM hyperspectral sensor is characterized by 192 bands ranging from 0.4 to 2.5 µm with a spatial resolution that may vary from 4 m to 8 m (Table 2). In order to simulate HYPXIM imagery, an airborne hyperspectral dataset was acquired using the HySpex sensor (Table 2) in July 2015. These data have been resampled accordingly to HYPXIM specificities at 4 m and 8 m spatial resolution. Moreover, to highlight the interest of VHSR hyperspectral images, a Sentinel2 multispectral dataset has also been simulated at 10m resolution (for all spectral bands). Atmospheric corrections have been applied using the Cochise (atmospheric COrrection Code for Hyperspectral Images of remote-sensing SEnsors) method [50]. This method assumes a flat ground hypothesis and estimates the water vapor content using a linear regression report (LIRR). It is particularly well suited for hyperspectral imagery [51].

2.3. In Situ Samples

Field work was conducted between May 31st and 2nd June 2017. It is assumed that between the data acquisition (2015) and field survey (2017), the tree composition of the city did not change. Indeed, trees are nowadays part of the climate adaptation and mitigation plan of the municipality, declared as protected contributors to cooling urban heat islands and summer heat waves. However, in the case of disease, some removals may occur, but to our knowledge, there were none during this period. This fieldwork led to the identification and location, using a differential GPS (Trimble Geo7x–with an absolute geometric accuracy below 10 cm), of more than 600 trees in the swath of the hyperspectral dataset. This accuracy allows for the extraction of each sample the corresponding spectral signature. Moreover, each inventoried tree—hereafter “individual”—has been classified into types, family, genus and species. A hierarchical nomenclature of 24 families and 43 species was established. This dataset is very unbalanced, e.g., Platanus hispanica (London plane) has 144 individuals while other species like Prunus avium (Cherry) has only seven individuals.

2.4. Methods

2.4.1. Overall Approach

The aim of this study is to assess the influence of hyperspectral techniques and data on their ability to classify tree diversity. Figure 2 sums up the main hyperspectral techniques used in this study. Several methods of DR (Section 2.4.2.) and classification (Section 2.4.3.) are tested to evaluate which combination of techniques offers the best performance according to the considered hyperspectral images. Each classification performance is assessed using various training size datasets (depending on tree Species and Families nomenclatures, themselves divided into two sub-levels according to the number of available individual samples per class inventoried in the field) and various sizes of stratified k-folds (see Section 2.4.3).

2.4.2. Dimension Reductions Methods

Feature extraction is defined as a set of methods for extracting information from an image [33], for example, by the use of transformative methods (PCA, MNF). PCA constructs a low-dimensional representation of the data that describes as much of the variance in the data as possible. This is done by finding a linear basis of reduced dimensionality for the data, in which the amount of variance in the data is maximal [34]. However, it does not take account of the signal-to-noise ratio (SNR) unlike the minimal noise fraction (MNF) [37]. The MNF is also a component transformation providing ordered noise-reduced components. Feature extraction methods (PCA, MNF) are applied on the four image datasets (HySpex 2 m, HYPXIM 4 m and 8 m, Sentine l2). The selection of the most appropriate components is made empirically, since they explain more than 98% of variance.
For feature selection methods, the literature brings a lot of information on discriminating indices nevertheless none has been applied in urban contexts. In this study, 96 spectral vegetation indices (VI) (listed in [37]) have been computed for the three hyperspectral datasets as Sentinel spectral resolution is not sufficient to do so. These VI illustrate various biophysical characteristics such as chlorophyll, leaf area index (LAI), carotenoids, etc., and the selection of the most uncorrelated VI is based on two methods:
  • The K-best method uses the F-value of ANOVA (analysis of variance) as a ranking criterion. It compares the interclass and intra-class dispersion between VI. Subsequently, from a correlation matrix (Pearson’s r), if the correlation between two VI is greater than 85% (r > 0.85) then the VI with the smallest F-value is removed from the selection.
  • A PCA method is used to eliminate the correlation between VI. Kaiser’s criterion [39] is used for the selection: components with a eigenvalue greater than 1 represents the same amount of information as a single variable.
Finally, DR methods provide 14 reduced hyperspectral datasets which will be compared with original (unreduced) hyperspectral and Sentinel2 datasets to assess their influence.

2.4.3. Training Datasets and Method

Two levels of tree nomenclature are considered: Families and Species. Types have been discarded because of their too general level, the unbalanced dataset but also because multispectral data are known to correctly classify this level of nomenclature. Families and species are respectively sub-divided into two training datasets, depending on the number of individuals per class: the first ones are less restrictive with at least four individuals, while the others consider at least 10 individuals per class. The number of tree classes considered depends on the representativeness of training data (Table 3).
The training method is based on stratified k-folds [44,47]: the original dataset is randomly partitioned into k subsamples where each class is represented with the same proportion of individuals. To assess the influence of the training size, two approaches are considered. The first one uses k-folds = 2 (2-folds): the dataset is divided into 50% training, 50% validation samples and applied for the four training datasets. More precisely, for each class, half of the individuals are used to extract the spectral signature (or assimilated values of reduced datasets) to train the classifier, and the other half is used to validate the resulting classification. In order to optimize the learning capacity of classifiers, each fold is used iteratively as a training and validation sample. The second one is similar to the first one and uses k-folds = 10 (10-folds). It is implemented only for training datasets with at least 10 individuals per class (“12 Families” and “14 Species”).

2.4.4. Classification Methods

Two commonly used, but not the only classification methods for hyperspectral data are tested, namely the support vector machine (SVM) and random forest (RF). The SVM projects the original feature space into a space with a higher dimensionality to facilitate the discrimination of classes. The classifier is trained using an optimization process associated with a cost function that makes it possible to linearly separate classes to kernel functions (linear, radial basis function—RBF—polynomial, sigmoidal) applied on the original data allowing greater reliability and classification accuracy [41]. However, it can sometimes lead to an over-fitted model with misclassification [41]. The RF classifier is an ensemble of many decision trees. This classifier is also known to be appropriate for hyperspectral data [42,43,52,53] because it tends to perform a selection of characteristics to build the decision tree during the learning process [43]. Its main advantage relies in the ability to stem over-fitting if the number of forests is sufficient. Increasing the predictive power of the model is performed here with an empirical optimization of three hyper-parameters: the number of forests, the maximum number of features, and the splitting criterion (Gini impurity or entropy for the information gain). For both the SVM and RF, hyperparameter optimization has been grid searched thanks to the available training datasets.

2.4.5. Experiments: Assessing the Respective and Combined Influence of Hyperspectral Techniques and Data

The respective influence of hyperspectral techniques and data is evaluated based on their ability to classify tree diversity. Classification accuracy is performed using the overall accuracy (OA) and the Kappa index of agreement (KIA). For the 2-folds learning method and data, each SVM and RF classifier is applied 10 times on each reduced and unreduced image dataset in order to estimate the variance of the OA, leading to the computation of 1120 urban tree classifications. Although these results are appropriate to assess the combined influence of the hyperspectral techniques, their respective influence is difficult to evaluate. We assume that averaged OA and KIA (AOA / AKIA) for each technique computed from these classifications can be useful to evaluate their respective contribution. Moreover, the influence of training methods (i.e., the number of k-folds) is assessed differently as the 10-folds is computationally highly time-consuming. It is applied only on the image dataset exhibiting the best classification results, and then compared with the corresponding results obtained with the 2-folds training method.
The combined effect of hyperspectral techniques is analyzed more precisely thanks to the OA and KIA, accounting for both the producer’s and user’s accuracy.

3. Results

3.1. Dimension Reductions: Comparison of Extracted Components

The number of selected components from MNF and PCA is respectively of 40 and 30 for all hyperspectral images, and 13 and 9 for Sentinel2. The feature selection method using PCA provides a selection of 19 vegetation index for HySpex 2 m (explaining 86.16% of the variance), 19 for HYPXIM 4 m (84.5%) and 17 for HYPXIM 8 m (85.89%). When using the ANOVA and the correlation matrix, 45 uncorrelated VI are selected for HySpex 2 m, 39 for HYPXIM 4 m and 38 for HYPXIM 8 m. A recurrence of selected VI is observed between datasets. For example the VI Datt (850) is selected for all image datasets. HYPXIM 4 m exhibit the larger F-values interval (between 38.26 and 1.96) compared to HySpex (from 34.26 and 0.06) and HYPXIM 8 m (27.96 and 1.73). The larger is the interval (i.e. the variance between selected VI), the greater is the separability power. The most selected VI regardless to the resolution rely to the chlorophyll content with 24/45 selected indices for HySpex, 19/39 for HYPXIM 4 m and 24/38 for HYPXIM 8 m. In addition, the first 10 VI exhibiting the highest F-values are also related to chlorophyll (except "CaCol_515_550" for the 8 m HYPXIM image that is linked to the carotenoid) meaning that this biophysical property has the most discriminating potential (cf. Supplementary Figure S1 for details).

3.2. Evaluation of the Respective Influence of Hyperspectral Techniques and Data

When using the 2-folds training method, both classifiers offer similar results. Classifications made with RF presents an AOA = 45.15% and an AOA = 42.61% with SVM (Table 4). The latter has the lower AOA while exhibiting the highest score (AOA = 59% and best AOA = 58.54% for RF). RF shows the greatest variance with on average σ = 4.2% against σ = 3.8% for SVM.
For the two considered classifiers, results obtained from the MNF image (with AOA = 55.09% and 55.91% for RF and SVM respectively—Table 4) datasets are always greater than those obtained from other reduced/unreduced image datasets. VI by ANOVA has an AOA of 45.83% (AOA = 49.65% and 49.65% for RF and SVM respectively) and VI by PCA 40.97% (AOA = 39.48% and 42.46% for RF and SVM respectively). PCA and unreduced image datasets show similar AOA (38.50% and 38.60% respectively).
Sub-level training datasets show that those with at least 10 samples (12 Families and 14 Species) help to improve tree classification accuracy compared to those with a minimal number of 4 individual samples. The sub-level “14 species” gives the best results with an AOA = 47.10% against 45.62% for the sub-level “12 families”. The sub-level “19 families” (n individuals > 4) produces results slightly higher than the sublevel “27 species” with respectively AOA = 42.23% and 40.56%. The comparison of the 2-folds and 10-folds method as the training method shows the potential of 10-folds method because it significantly improves AOA. When comparing results obtained with these two training methods applied on MNF image datasets, the 10-folds provides an AOA = 68.26% although AOA = 58.42% with the 2-folds one (Table 5 and Table 6).
Finally, according to the spatial and spectral resolution, HySpex 2 m show an AOA = 52.60% that is greater than any others image datasets: AOA = 42.38% for HYPXIM 4 m, AOA = 40.51% for HYPXIM 8 m and AOA = 41.93% for Sentinel2 10m (Table 7).

3.3. Influence of Combined Techniques According to Images

The best tree diversity classification accuracy (OA = 78.38%) is obtained using the reduced image HYPXIM 4 m by MNF, with the sub-level training sample “14 species” (n individuals > 10), the 10-folds training method and with SVM classifier. A similar score is obtained with HySpex 2 m while applying the same hyperspectral techniques and training data and method. KIA are also similar for the both reduced image (KIA = 0.73) but reduced HySpex present a greater variance than reduced HYPXIM at 4 m (σ = 2.5% and σ = 1.5% respectively) (Table 5 and Figure 3). The “12 families” nomenclature produces also good results with OA = 73.81% and KIA = 0.68 with reduced images HySpex and HYPXIM at 4 m however the latter presents a smaller standard deviation (±5%). HYPXIM 8 m and Sentinel2 show a similar OA = 61.90% (Table 5 and Figure 4).
Although the 10-folds always provides higher results compared to 2-folds learning method (with MNF DR only), it reinforces the interest of HYPXIM imageries. Indeed, with the 10-folds learning method, a significant improvement of urban tree diversity classification accuracy is observed: HySpex 2 m and HYPXIM 4 m reduced with MNF both exhibit an OA of 78.38% and 73.81% for 12 families and 14 species classes, respectively (Figure 3 and Figure 4). In the meanwhile, improvements for HYPXIM 8 m and Sentinel2 due to 10-folds are much lower, or exhibit an OA below 65%.
Focusing on the classification images exhibiting the highest OA (Figure 5 and Figure 6), helps to understand classification errors in the context of the obtained scores. Indeed, for the reduced HySpex image, an overestimation of London plane is visible with 2-folds method (Figure 5.1.a), as well as for Linden sp. with 10-folds method (Figure 5.2.a). In both cases, London plane is correctly detected on large boulevard (Bd. Lazare Carnot) while over-estimated due to confusion with the shadow of buildings. In specific areas exhibiting a great tree diversity such as the public park of the “Grand-Rond” roundabout, the 10-folds method (Figure 5.2.b) gives a better prediction than 2-folds (Figure 5.1.b) in particular for Silver linden that are correctly detected. Conversely, it contributes to misclassify neighboring pixels of tree alignment on the previously cited boulevard north to the Grand Rond (Figure 5.2.b): Linden are over estimated although only London plane species is present. These misclassifications are less present when classification is applied with reduced HYPXIM image at 4 m (Figure 6). Species are better detected with less errors of neighboring pixels (Figure 6(1.b,2.b)). However, the classification with 2-folds method (Figure 6.1.b) evinces a bigger misclassification than 10-folds method as for example on the Grand-Rond roundabout where Silver linden are over predicted.
Related confusion matrices (Supplementary Table S2) and producer’s and user’s accuracies (respectively PA and UA – Supplementary Table S3) highlight the influence of the unbalanced in situ dataset. Table 8 illustrates the PA and UA for the classifications shown in Figure 5 and Figure 6.
First, the greater the number of individuals per tree species, the better the PA and UA. Then, UA are in most of the cases greater than the PA, except for the well represented tree species. Table S2 shows that tree species with a low PA are in most of the cases mistakenly classified as London plane. Third, the number of k-folds significantly improves accuracies for under-represented tree species, particularly for MNF-reduced HySpex images. Finally, the number of missed tree species (0% of PA or UA) is slightly greater with RF than SVM for both 14 species’ and 12 families’ nomenclatures.

4. Discussion

4.1. Evaluating the Respective Influence of Hyperspectral Techniques

First of all, comparing the respective influence of the hyperspectral techniques considered remains a difficult task since the influence of one technique on another is not linear. Although the use of averaged OA values per technique can be criticized, this provides interesting information on the respective contribution of the methods that have been tested for the classification of tree diversity. The method that allows the greatest gain in overall accuracy depends on the learning process and the data. In our case, the gain can be greater than 15% for some image datasets (HySpex 2 m, HYPXIM 4 m) as soon as the number of k-folds with a minimum of 10 individuals per class increases. Second, dimensional reduction methods also improve the classification’s overall accuracy than other classification methods. However, depending on the considered image dataset or even the DR methods for instance, some may be punctually more effective than others, fuelling the debate on the justification for choosing one or the other method. Nevertheless, it can be assumed that the very high spatial resolution of hyperspectral data has an inherent limitation: it increases the dimensionality of the data and makes any classification method ineffective if dimensional reduction methods are not previously applied. Indeed, in some few cases, the Sentinel2 image gives similar or even slightly better results than hyperspectral images even though the overall accuracy remains quite low.
While we have focused on dimensional reduction methods, our results are in line with the scientific literature. The most effective method is MNF. The selection of vegetation indices gives interesting results and better averages than the use of unreduced hyperspectral datasets. The MNF’s ability to reduce noise while minimizing signal loss explains this level of performance unlike the ACP. Feature selection methods remain less accurate than feature extraction ones. While considering the ANOVA and correlation matrix selection method, low results can be explained by empirically choosing the correlation threshold. Oldeland et al. [54] used a threshold of r = 0.75 without any explanations. Another selection method could be tested such as the Kruskal Wallis algorithm which iteratively tests the best combination of uncorrelated indices [37]. Moreover, these authors have shown that not only vegetation indices can be used, but also various derivatives (1st order or 2nd order) of (part of) the spectral signature. Indeed, derivatives are insensible to variations in illumination intensity caused by changes in sun angle [55,56] and which may be affected by the floristic composition and vegetation type [37].

4.2. Characterizing Urban Tree Diversity with VHRS Hyperspectral Remote Sensing: Advantages, Limitations and Prospects

This study presents results similar to those described in the literature cited in the introduction. Thanks to optimized hyperspectral data and techniques, about 80% of urban tree diversity (14 different species) can be accurately discriminated against. These are fairly good results given that hyperspectral data were acquired in summer, which might not be the optimal season to distinguish species in comparison to late spring or fall.
The spatial and spectral resolutions of HYPXIM images present great potential for such an application. Indeed, in general, the results decrease with the decrease in spatial and spectral resolutions. In some cases, Sentinel2 images give better results than those obtained with HYPXIM. But as soon as hyperspectral techniques are optimized, there is a real qualitative leap allowing HYPXIM 4 m data to obtain results similar to those obtained with HySpex 2 m data. This suggests that 4 m can be an almost optimal spatial resolution for classifying urban tree diversity. On the one hand, at 2 m, even if each tree can be better delineated, classification errors may increase due to the variability of spectral signatures in a crown [48]. Indeed, spectral reflectance can vary within the same tree, depending on biochemistry and water content [57,58], canopy architecture, woody biomass exposure and underlying substrate exposure [30]. Figure 5 and Figure 6 show this effect by comparing classifications from HySpex 2 m and HYPXIM 4 m. North of the ‘Grand-Rond’ roundabout, the boulevard is entirely composed of London plane while between two rows of trees some pixels are assimilated to Linden trees with HySpex images. These errors are rare with HYPXIM images at 4 m. On the other hand, a coarser spatial resolution does not help because of mixed pixels [40]. However, spectral resolution also influences the ability to distinguish urban tree species with similar spectral signatures [57]: the choice of spectral resolution depends on how detailed the nomenclature is.
The purpose of this study was to evaluate the capabilities of a future hyperspectral spaceborne sensor (HYPXIM) for a specific urban environmental application (urban tree classification). Given the satellite’s revisiting capacity, it can be assumed that VHRS hyperspectral imaging time series would greatly improve the classification of urban tree diversity through phenological differentiation [18] while considering the development of new methods to cope with the increase size of such a data set. Recently, others innovative methods have proven their relevance for hyperspectral data and techniques: object-based or textural information are relevant for detecting vegetation types [59]; deep-learning methods and other neural networks efficiently addressed the large dimensionality of hyperspectral images for classification or dimensional reduction [60,61].

4.3. Towards an Urban Sampling Strategy for Mapping Urban Tree Diversity?

This study shows that the accuracy of urban tree diversity classification strongly depends on the field dataset which in turn depends on the defined nomenclature. Considering the producer’s and user’s accuracy of tree classification (c.f. Supplementary Table S2), it is advisable to define in advance an appropriate nomenclature based on a discriminating analysis of species and families classes. Indeed, some family classes can be better separated thanks to their spectral signature than two different species, or inversely. For instance, Tilia tomentosa (Silver linden) and Tilia disticum (Linden) species are well distinguished while belonging to the same family. Spectra variability can vary according to many factors inherited from urban morphology and urban tree management: street width and orientation, the shading effects of buildings or nearby species, tree age and size [57,58], etc.
Another limitation is the unbalanced data set that is representative of the distribution of trees generally found in urban areas. In the past, urban vegetation plans often explained why a few species dominate in cities. The diversity of trees is also due to individual initiatives. However, if a balanced sampling strategy can be set to minimize their influence on the classification accuracy, their precise geographical identification may be difficult because private yards are not easily accessible. In this study, the Platanaceae family is represented by a single species (London plane), which is not the case for other families. This can affect the classification accuracy as some classes are over-represented, resulting in an underestimation of Silver Linden and misclassified Linden species.
However, hyperspectral data have limited classification capabilities with a limited number of training samples [62] due to possible over-adjustment to learning data [63]. Increasing the -folds in the learning process can help to minimize the influence of the unbalanced sampling by preventing from overfitting. Although further studies are needed to define an optimal sampling strategy for urban areas, we assume this strategy must be adapted to the pursued objective: Do we want to have a good map of all the species/families in a considered city? Do we want to map the dominant/specific species? While the in situ dataset should be well defined at the beginning of the study, other techniques may also be of interest to evaluate. For example, object-based image analysis (OBIA) which aim is to segment an image based not only on averaged or pixel majority spectral signatures but also on their spatial arrangement [64,65,66,67], can be helpful in minimizing neighboring misclassifications by reducing pixel-level noise while combining spectrally similar pixels.

5. Conclusions

This research aimed to assess the ability of VHSR hyperspectral satellite data in order to discriminate against urban tree diversity. The influence of hyperspectral techniques (dimension reduction and classification methods as well as learning data and techniques) on the classification accuracy was also evaluated on the case study of Toulouse, based on two sublevels of taxonomic nomenclature (species and families) with a minimal number of individual samples per class (4 to 10). All datasets are available upon request for the moment and will be freely available once published.
This study highlighted the great potential of VHRS (4 m) hyperspectral images (192 spectral bands). Combined with MNF dimension reduction method, the SVM classifier and a 10-folds learning method with 10 individuals per class, the discrimination of 14 species reached an overall accuracy of 78.4% (KIA = 0.73) which is consistent with the existing literature. This performance is equal to that obtained with HySpex images (2 m, 408 spectral bands). By comparing the influence of each technique, it shows that the most contributory rely upon learning methods and data. Dimension reduction methods have also a significant influence on the classification accuracy: the MNF feature extraction method is the most efficient compared to feature selection methods. The selection of vegetation indices has shown interesting potential but requires further developments to optimize their selection.
Finally, although this study was conducted using simulated spaceborne data from airborne hyperspectral images for a single, not optimal date, spaceborne VHRS hyperspectral images would help to take a major step forward by providing even more consistent time series in order to distinguish urban tree diversity through their phenology. However, urban areas have inherent limitations due to the very diverse and unbalanced composition of the urban tree canopy. Further work is needed to evaluate and define a sampling strategy adapted to urban areas.

Supplementary Materials

The following are available online at https://0-www-mdpi-com.brum.beds.ac.uk/2072-4292/11/11/1269/s1, Figure S1: Results of ANOVA for the selected Vegetation indices applied on hyperspectral datasets; Table S2: Confusion matrices for the SVM classification applied with 10-Folds method on (top) reduced image HySpex at 2 m by MNF and (down) reduced image HYPXIM 4 m by MNF; Table S3: Tables showing the Producer’s and User’s Accuracies (PA and UA) obtained from the MNF-reduced image datasets, with RF and SVM classifiers trained with 2-folds and 10-folds for the 14 species training dataset, on the one hand, and the 12 families on the other hand.

Author Contributions

Conceptualization: T.H., A.L. and C.B.; methodology: T.H. and C.B.; software: E.A.-V., C.B. and A.T.; validation: E.A.-V. and C.B.; formal analysis: C.B., E.A.-V., A.L., G.M., K.T.N.; investigation: C.B., E.A.-V., A.L., G.M., K.T.N.; resources: E.A.-V., C.B. and A.T.; data curation: C.B., E.A.-V., A.L., G.M., K.T.N., X.X.; Writing—Original Draft preparation: C.B.; Writing—Review and Editing: T.H.; supervision: T.H.; project administration: T.H.; funding acquisition: T.H.

Funding

This research was funded by the French National Science foundation project HYEP (Hyperspectral imagery for environmental Urban planning). Project funded by the ANR 14-CE22-0016-01.

Acknowledgments

Thanks to Thomas Corpetti for his technical support and help for the implementation of classifications. Authors would like to thank the two anonymous reviewers for their helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Masson, V.; Marchadier, C.; Adolphe, L.; Aguejdad, R.; Avner, P.; Bonhomme, M.; Bretagne, G.; Briottet, X.; Bueno, B.; de Munck, C.; et al. Adapting cities to climate change: A systemic modelling approach. Urban Clim. 2014, 10, 407–429. [Google Scholar] [CrossRef]
  2. Alberti, M. The Effects of Urban Patterns on Ecosystem Function. Int. Reg. Sci. Rev. 2005, 28, 168–192. [Google Scholar] [CrossRef]
  3. El Araby, M. Urban growth and environmental degradation. Cities 2002, 19, 389–400. [Google Scholar] [CrossRef]
  4. Pickett, S.T.; Burch, W.R., Jr.; Dalton, S.E.; Foresman, T.W.; Grove, J.M.; Rowntree, R. A conceptual framework for the study of human ecosystems in urban areas. Urban Ecosyst. 1997, 185–199. [Google Scholar] [CrossRef]
  5. Shepherd, J.M. A Review of Current Investigations of Urban-Induced Rainfall and Recommendations for the Future. Earth Interact. 2005, 9, 1–27. [Google Scholar] [CrossRef] [Green Version]
  6. Escobedo, F.J.; Nowak, D.J. Spatial heterogeneity and air pollution removal by an urban forest. Landsc. Urban Plan. 2009, 90, 102–110. [Google Scholar] [CrossRef]
  7. Akbari, H.; Pomerantz, M.; Taha, H. Cool surfaces and shade trees to reduce energy use and improve air quality in urban areas. Sol. Energy 2001, 70, 295–310. [Google Scholar] [CrossRef]
  8. Bolund, P.; Hunhammar, S. Ecosystem services in urban areas. Ecol. Econ. 1999, 29, 293–301. [Google Scholar] [CrossRef]
  9. Santamouris, M. Using cool pavements as a mitigation strategy to fight urban heat island—A review of the actual developments. Renew. Sustain. Energy Rev. 2013, 26, 224–240. [Google Scholar] [CrossRef]
  10. McCarthy, H.R.; Pataki, D.E. Drivers of variability in water use of native and non-native urban trees in the greater Los Angeles area. Urban Ecosyst. 2010, 13, 393–414. [Google Scholar] [CrossRef] [Green Version]
  11. McPherson, E.G.; Simpson, J.R.; Xiao, Q.; Wu, C. Million trees Los Angeles canopy cover and benefit assessment. Landsc. Urban Plan. 2011, 99, 40–50. [Google Scholar] [CrossRef]
  12. Simpson, J.R. Improved estimates of tree-shade effects on residential energy use. Energy Build. 2002, 34, 1067–1076. [Google Scholar] [CrossRef]
  13. Lemonsu, A.; Masson, V.; Shashua-Bar, L.; Erell, E.; Pearlmutter, D. Inclusion of vegetation in the Town Energy Balance model for modelling urban green areas. Geosci. Model Dev. 2012, 5, 1377–1393. [Google Scholar] [CrossRef] [Green Version]
  14. Redon, E.C.; Lemonsu, A.; Masson, V.; Morille, B.; Musy, M. Implementation of street trees within the solar radiative exchange parameterization of TEB in SURFEX v8.0. Geosci. Model Dev. 2017, 10, 385–411. [Google Scholar] [CrossRef] [Green Version]
  15. Pu, R.; Landry, S.; Yu, Q. Assessing the potential of multi-seasonal high resolution Pléiades satellite imagery for mapping urban tree species. Int. J. Appl. Earth Obs. Geoinf. 2018, 71, 144–158. [Google Scholar] [CrossRef]
  16. Pu, R.; Landry, S. A comparative analysis of high spatial resolution IKONOS and WorldView-2 imagery for mapping urban tree species. Remote Sens. Environ. 2012, 124, 516–533. [Google Scholar] [CrossRef]
  17. Alonzo, M.; Bookhagen, B.; Roberts, D.A. Urban tree species mapping using hyperspectral and lidar data fusion. Remote Sens. Environ. 2014, 148, 70–83. [Google Scholar] [CrossRef]
  18. Voss, M.; Sugumaran, R. Seasonal Effect on Tree Species Classification in an Urban Environment Using Hyperspectral Data, LiDAR, and an Object-Oriented Approach. Sensors 2008, 8, 3020–3036. [Google Scholar] [CrossRef] [PubMed]
  19. Xiao, Q.; Ustin, S.L.; McPherson, E.G. Using AVIRIS data and multiple-masking techniques to map urban forest tree species. Int. J. Remote Sens. 2004, 25, 5637–5654. [Google Scholar] [CrossRef] [Green Version]
  20. Boschetti, M.; Boschetti, L.; Oliveri, S.; Casati, L.; Canova, I. Tree species mapping with Airborne hyper-spectral MIVIS data: the Ticino Park study case. Int. J. Remote Sens. 2007, 28, 1251–1261. [Google Scholar] [CrossRef]
  21. Maschler, J.; Atzberger, C.; Immitzer, M. Individual Tree Crown Segmentation and Classification of 13 Tree Species Using Airborne Hyperspectral Data. Remote Sens. 2018, 10, 1218. [Google Scholar] [CrossRef]
  22. Immitzer, M.; Vuolo, F.; Atzberger, C. First Experience with Sentinel-2 Data for Crop and Tree Species Classifications in Central Europe. Remote Sens. 2016, 8, 166. [Google Scholar] [CrossRef]
  23. Gaia Vaglio, L.; Pulleti, N.; Hawthorne, W.; Liesenberg, V.; Corona, P.; Papale, D.; Chen, Q.; Valentini, R. Discrimination of tropical forest types, dominant species, and mapping of functional guilds by hyperspectral and simulated multispectral Sentinel-2 data. Remote Sens. Environ. 2016, 176, 163–176. [Google Scholar] [CrossRef] [Green Version]
  24. Shi, Y.; Skidmore, A.K.; Wang, T.; Holzwarth, S.; Heiden, U.; Pinnel, N.; Zhu, X.; Heurich, M. Tree species classification using plant functional traits from LiDAR and hyperspectral data. Int. J. Appl. Earth Obs. Geoinf. 2018, 73, 207–219. [Google Scholar] [CrossRef]
  25. Mozgeris, G.; Juodkienė, V.; Jonikavičius, D.; Straigytė, L.; Gadal, S.; Ouerghemmi, W. Ultra-Light Aircraft-Based Hyperspectral and Colour-Infrared Imaging to Identify Deciduous Tree Species in an Urban Environment. Remote Sens. 2018, 10, 1668. [Google Scholar] [CrossRef]
  26. Dabiri, Z.; Lang, S. Comparison of Independent Component Analysis, Principal Component Analysis, and Minimum Noise Fraction Transformation for Tree Species Classification Using APEX Hyperspectral Imagery. ISPRS Int. J. Geo Inf. 2018, 7, 488. [Google Scholar] [CrossRef]
  27. Liu, H.; Wu, C. Crown-level tree species classification from AISA hyperspectral imagery using an innovative pixel-weighting approach. Int. J. Appl. Earth Obs. Geoinf. 2018, 68, 298–307. [Google Scholar] [CrossRef]
  28. Dalponte, M.; Bruzzone, L.; Vescovo, L.; Gianelle, D. The role of spectral resolution and classifier complexity in the analysis of hyperspectral images of forest areas. Remote Sens. Environ. 2009, 113, 2345–2355. [Google Scholar] [CrossRef]
  29. Noyel, G. Filtrage, Réduction de Dimension, Classification et Segmentation Morphologique Hyperspectrale. Ph.D. Thesis, Mines ParisTech, University PSL, Paris, France, 2008; 280p. [Google Scholar]
  30. Landgrebe, D.A. Signal Theory Methods in Multispectral Remote Sensing; John Wiley & Sons: Hoboken, NJ, USA, 2005; Volume 29, ISBN 978-0-471-72125-3. [Google Scholar]
  31. Lu, D.; Weng, Q. A survey of image classification methods and techniques for improving classification performance. Int. J. Remote Sens. 2007, 28, 823–870. [Google Scholar] [CrossRef]
  32. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  33. Girard, M.-C.; Girard, C.M. Traitement des Données de Télédétection: Environnement et Ressources Naturelles, 2nd ed.; Technique et Ingénierie; Dunod: Paris, France, 2017; ISBN 2-10-076370-9. [Google Scholar]
  34. Khoder, J.; Younes, R. Dimensionality reduction on hyperspectral images: A comparative review based on artificial datas. In Proceedings of the 4th International Congress on Image and Signal Processing, Shanghai, China, 15–17 October 2011; pp. 1875–1883. [Google Scholar]
  35. Farrell, M.D.; Mersereau, R.M. On the Impact of PCA Dimension Reduction for Hyperspectral Detection of Difficult Targets. IEEE Geosci. Remote Sens. Lett. 2005, 2, 192–195. [Google Scholar] [CrossRef]
  36. Green, A.A.; Berman, M.; Switzer, P. A Transformation for Ordering Multispectral Data in Terms of Image Quality with Implications for Noise Removal. IEEE Trans. Geosci. Remote Sens. 1988, 26, 65–74. [Google Scholar] [CrossRef]
  37. Erudel, T.; Fabre, S.; Houet, T.; Mazier, F.; Briottet, X. Criteria Comparison for Classifying Peatland Vegetation Types Using In Situ Hyperspectral Measurements. Remote Sens. 2017, 9, 748. [Google Scholar] [CrossRef]
  38. Clark, M.; Roberts, D.; Clark, D. Hyperspectral discrimination of tropical rain forest tree species at leaf to crown scales. Remote Sens. Environ. 2005, 96, 375–398. [Google Scholar] [CrossRef]
  39. Kaiser, H.F. The Application of Electronic Computers to Factor Analysis. Educ. Psychol. Meas. 1960, 20, 141–151. [Google Scholar] [CrossRef]
  40. Becker, B.L.; Lusch, D.P.; Qi, J. A classification-based assessment of the optimal spectral and spatial resolutions for Great Lakes coastal wetland imagery. Remote Sens. Environ. 2007, 108, 111–120. [Google Scholar] [CrossRef]
  41. Camps-Valls, G.; Bruzzone, L. Kernel Methods for Remote Sensing Data Analysis; Wiley: Hoboken, NJ, USA, 2009; p. 15. [Google Scholar]
  42. Chan, J.C.-W.; Paelinckx, D. Evaluation of Random Forest and Adaboost tree-based ensemble classification and spectral band selection for ecotope mapping using airborne hyperspectral imagery. Remote Sens. Environ. 2008, 112, 2999–3011. [Google Scholar] [CrossRef]
  43. Dalponte, M.; Orka, H.O.; Gobakken, T.; Gianelle, D.; Naesset, E. Tree Species Classification in Boreal Forests with Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2632–2645. [Google Scholar] [CrossRef]
  44. Foody, G.M.; Mathur, A. Toward intelligent training of supervised image classifications: Directing training data acquisition for SVM classification. Remote Sens. Environ. 2004, 93, 107–117. [Google Scholar] [CrossRef]
  45. Plaza, A.; Benediktsson, J.A.; Boardman, J.W.; Brazile, J.; Bruzzone, L.; Camps-Valls, G.; Chanussot, J.; Fauvel, M.; Gamba, P.; Gualtieri, A.; et al. Recent advances in techniques for hyperspectral image processing. Remote Sens. Environ. 2009, 113, S110–S122. [Google Scholar] [CrossRef]
  46. Willis, C.J. Hyperspectral Image Classification with Limited Training Data Samples Using Feature Subspaces. In Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery X; Shen, S.S., Lewis, P.E., Eds.; SPIE: Bellingham, WA, USA, 2004; p. 170. [Google Scholar]
  47. Rodriguez, J.D.; Perez, A.; Lozano, J.A. Sensitivity Analysis of k-Fold Cross Validation in Prediction Error Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 569–575. [Google Scholar] [CrossRef] [PubMed]
  48. Alonzo, M.; Roth, K.; Roberts, D. Identifying Santa Barbara’s urban tree species from AVIRIS imagery using canonical discriminant analysis. Remote Sens. Lett. 2013, 4, 513–521. [Google Scholar] [CrossRef]
  49. Karoui, M.; Benhalouche, F.; Deville, Y.; Djerriri, K.; Briottet, X.; Le Bris, A. Detection and area estimation for photovoltaic panels in urban hyperspectral remote sensing data by an original NMF-based unmixing method. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS 2018), Valencia, Spain, 22–27 July 2018. [Google Scholar]
  50. Miesch, C.; Poutier, L.; Achard, V.; Briottet, X.; Lenot, X.; Boucher, Y. Direct and Inverse Radiative Transfer Solutions for Visible and Near-Infrared Hyperspectral Imagery. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1552–1562. [Google Scholar] [CrossRef]
  51. Roussel, G.; Weber, C.; Briottet, X.; Ceamanos, X. Comparison of two atmospheric correction methods for the classification of spaceborne urban hyperspectral data depending on the spatial resolution. Int. J. Remote Sens. 2017, 39, 1593–1614. [Google Scholar] [CrossRef]
  52. Crawford, M.M.; Ham, J.; Chen, Y.; Ghosh, J. Random forests of binary hierarchical classifiers for analysis of hyperspectral data. In Proceedings of the IEEE Workshop on Advances in Techniques for Analysis of Remotely Sensed Data, Greenbelt, MD, USA, 27–28 October 2003; pp. 337–345. [Google Scholar]
  53. Ham, J.; Chen, Y.; Crawford, M.M.; Ghosh, J. Investigation of the random forest framework for classification of hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2005, 43, 492–501. [Google Scholar] [CrossRef] [Green Version]
  54. Oldeland, J.; Dorigo, W.; Lieckfeld, L.; Lucieer, A.; Jürgens, N. Combining vegetation indices, constrained ordination and fuzzy classification for mapping semi-natural vegetation units from hyperspectral imagery. Remote Sens. Environ. 2010, 114, 1155–1166. [Google Scholar] [CrossRef]
  55. Feilhauer, H.; Asner, G.P.; Martin, R.E.; Schmidtlein, S. Brightness-normalized Partial Least Squares Regression for hyperspectral data. J. Quant. Spectrosc. Radiat. Transf. 2010, 111, 1947–1957. [Google Scholar] [CrossRef]
  56. Tsai, F.; Philpot, W. Derivative Analysis of Hyperspectral Data. Remote Sens. Environ. 1998, 66, 41–51. [Google Scholar] [CrossRef]
  57. Van Aardt, J.A.N.; Wynne, R.H. Examining pine spectral separability using hyperspectral data from an airborne sensor: An extension of field-based results. Int. J. Remote Sens. 2007, 28, 431–436. [Google Scholar] [CrossRef]
  58. Zhang, C.; Qiu, F. Mapping Individual Tree Species in an Urban Forest Using Airborne Lidar Data and Hyperspectral Imagery. Photogramm. Eng. Remote Sens. 2012, 78, 1079–1087. [Google Scholar] [CrossRef] [Green Version]
  59. Xie, Y.; Sha, Z.; Yu, M. Remote sensing imagery in vegetation mapping: A review. J. Plant Ecol. 2008, 1, 9–23. [Google Scholar] [CrossRef]
  60. Feng, Q.; Liu, J.; Gong, J. UAV remote sensing for urban vegetation mapping using random forest and texture analysis. Remote Sens. 2015, 7, 1074–1094. [Google Scholar] [CrossRef]
  61. Tadjudin, S.; Landgrebe, D.A. Covariance estimation with limited training samples. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2113–2118. [Google Scholar] [CrossRef] [Green Version]
  62. Chi, M.; Feng, R.; Bruzzone, L. Classification of hyperspectral remote-sensing data with primal SVM for small-sized training dataset problem. Adv. Space Res. 2008, 41, 1793–1799. [Google Scholar] [CrossRef]
  63. Benz, U.C.; Hofmann, P.; Willhauck, G.; Lingenfelder, I.; Heynen, M. Multi-resolution, object-oriented fuzzy analysis of remote sensing data for GIS-ready information. ISPRS J. Photogramm. Remote Sens. 2004, 58, 239–258. [Google Scholar] [CrossRef]
  64. Blaschke, T. Object based image analysis for remote sensing. ISPRS J. Photogramm. Remote Sens. 2010, 65, 2–16. [Google Scholar] [CrossRef] [Green Version]
  65. Myint, S.W.; Gober, P.; Brazel, A.; Grossman-Clarke, S.; Weng, Q. Per-pixel vs. object-based classification of urban land cover extraction using high spatial resolution imagery. Remote Sens. Environ. 2011, 115, 1145–1161. [Google Scholar] [CrossRef]
  66. Zhang, X.; Sun, Y.; Zhang, J.; Wu, P.; Jiao, L. Hyperspectral Unmixing via Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2018, 15, 1755–1759. [Google Scholar] [CrossRef]
  67. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
Figure 1. True color composite of hyperspectral and Sentinel2 images. (A.1) HySpex image at 2 m; (B.1) HYPXIM at 4 m; (C.1) HYPXIM at 8 m; (D.1) Sentinel2 at 10m. (A.2); (B.2); (C.2); (D.2): Zoom in the city center—RGB (µm): 0.665/0.560/0.490.
Figure 1. True color composite of hyperspectral and Sentinel2 images. (A.1) HySpex image at 2 m; (B.1) HYPXIM at 4 m; (C.1) HYPXIM at 8 m; (D.1) Sentinel2 at 10m. (A.2); (B.2); (C.2); (D.2): Zoom in the city center—RGB (µm): 0.665/0.560/0.490.
Remotesensing 11 01269 g001
Figure 2. General methodological approach. Box 1 lists the tested Dimension Reduction methods providing 14 reduced hyperspectral datasets and compared with the 3 unreduced original ones and the simulated Sentinel2 data—MNF and PCA are also applied on Sentinel2 (cf. Section 2.4.2). Box 2 lists the training data and methods used to train and validate the classification (cf. Section 2.4.3). Box 3 represents the two classifiers and criteria used to evaluate classifications’ accuracy (cf. Section 2.4.4).
Figure 2. General methodological approach. Box 1 lists the tested Dimension Reduction methods providing 14 reduced hyperspectral datasets and compared with the 3 unreduced original ones and the simulated Sentinel2 data—MNF and PCA are also applied on Sentinel2 (cf. Section 2.4.2). Box 2 lists the training data and methods used to train and validate the classification (cf. Section 2.4.3). Box 3 represents the two classifiers and criteria used to evaluate classifications’ accuracy (cf. Section 2.4.4).
Remotesensing 11 01269 g002
Figure 3. Comparison of results (OA) of 2-folds and 10-folds methods for the “species” nomenclature with the MNF DR.
Figure 3. Comparison of results (OA) of 2-folds and 10-folds methods for the “species” nomenclature with the MNF DR.
Remotesensing 11 01269 g003
Figure 4. Comparison of results (OA) of 2-fold and 10-folds methods for the “families” nomenclature with the MNF DR.
Figure 4. Comparison of results (OA) of 2-fold and 10-folds methods for the “families” nomenclature with the MNF DR.
Remotesensing 11 01269 g004
Figure 5. SVM classification of trees species (“14 Species” dataset) with MNF dimension reduction applied on HySpex at 2 m. (1.a) 2-folds method and (1.b) Zoom on Grand-Rond roundabout; (2.a) 10-folds method and (2.b) Zoom on Grand-Rond roundabout.
Figure 5. SVM classification of trees species (“14 Species” dataset) with MNF dimension reduction applied on HySpex at 2 m. (1.a) 2-folds method and (1.b) Zoom on Grand-Rond roundabout; (2.a) 10-folds method and (2.b) Zoom on Grand-Rond roundabout.
Remotesensing 11 01269 g005
Figure 6. SVM classification of trees species (“14 Species” dataset) with MNF dimension reduction applied on HYPXIM at 4 m. (1.a) 2-folds method and (1.b) Zoom on Grand-Rond roundabout; (2.a) 10-folds method and (2.b) Zoom on Grand-Rond roundabout.
Figure 6. SVM classification of trees species (“14 Species” dataset) with MNF dimension reduction applied on HYPXIM at 4 m. (1.a) 2-folds method and (1.b) Zoom on Grand-Rond roundabout; (2.a) 10-folds method and (2.b) Zoom on Grand-Rond roundabout.
Remotesensing 11 01269 g006
Table 1. State of this art of VHSR multispectral and hyperspectral images capabilities to discrimination of trees species in urban area.
Table 1. State of this art of VHSR multispectral and hyperspectral images capabilities to discrimination of trees species in urban area.
TypesSensorsSpatial ResolutionSpectral ResolutionOverall Accuracies/Number of Discriminated SpeciesStudy
MultispectralPleiades0.7 m-Panchromatic
2.4 m-Multispectral
463.51% for 7 species[15]
Ikonos1 m-Panchromatic
4 m-Multispectral
456.98% for 7 species[16]
Wordview 20.46 m-Panchromatic
1.84 m-Multispectral
862.93% for 7 species
HyperspectralAVIRIS3.5 m22479.2% for 29 species[17]
AISA2 m63Summer = 48% for 7 species
Fall = 45%
[18]
AVIRIS3.5 m13169% for 4 evergreen species
70% for 12 deciduous species
[19]
MIVIS4 m10275% for 7 families[20]
HySpex 0.4 m8091.7% for 13 species[21]
HySpex2 m29069.3% for 5 species[24]
Rikola0.7 m6463% for 6 species[25]
APEX2.5 m28890% for 7 species[26]
Hyperspectral + LiDARAVIRIS + LiDAR3.5 m22483.4% for 29 species[17]
AISA + LiDAR2 m63Summer = 57%
Fall = 56%
[18]
Hyspex + LiDAR2 m29083.7% for 5 species[24]
AISA + LiDAR1 m36682.12% for 4 species[27]
Table 2. Spectral and spatial characteristic of images used.
Table 2. Spectral and spatial characteristic of images used.
SensorSpatial ResolutionBandsDynamic RangeSpectral Resolution
HySpex2 m4080.4–2.5 µm0.41 to 0.96 µm = 3.64 nm
0.96 to 2.5 µm = 6 nm
HYPXIM4 m19210.9 nm
HYPXIM8 m192
Sentinel2 (simulated)10 m40.4–0.8 µm38–145 nm
60.7–2.2 µm18–242 nm
30.4–1.3 µm26–75 nm
Table 3. Training datasets according to families and species nomenclature of trees and their representativeness (number of individuals surveyed during the fieldwork).
Table 3. Training datasets according to families and species nomenclature of trees and their representativeness (number of individuals surveyed during the fieldwork).
FAMILIESSPECIES
Familiy NameNumber of IndividualsDatasetsSpecies NameNumber of IndividualsDatasets
Aceraceae4“19 families” Silver birch4“4 species”
Betulaceae4 Caucasian Wingnut4
Cupressaceae4 Pine sp4
Juglandaceae4 Wych elm4
Mimosaceae5 Caucasian Zelkova4
Rosaceae8 Albizia sp5
Salicaceae8 Atlas cedar5
Mixed shrub14“12 families”Deodar cedar5
Arecaceae16Cedar sp5
Lawn17Poplar sp5
Magnoliaceae21Western redbud6
Fagaceae22Cherry7
Oleaceae23Holm oak8
Fabaceae26Maple sp10 "10 species "
Sapindaceae29Oak12
Pinaceae32Fir sp13
Ulmaceae49Shrub mix14
Tiliaceae80Dwarf plam16
Platanaceae144Southern magnolia16
Number of tree classes1912Lawn17
Total of individuals510473White linden17
Horse chesnut21
number of individuals > 4Black locust21
number of individuals > 10Common privet22
Mediterranean hackberry35
Silver Linden63
London plane144
Number of tree classes2714
Total of individuals487421
Table 4. Urban tree diversity classification averaged overall accuracies (AOA) and averaged Kappa index of agreement (AKIA) obtained using the 2-folds learning method and applied on each reduced hyperspectral image, and on the original (hyperspectral and Sentinel2) image datasets with RF and SVM according to the four sub-levels for families and species datasets.
Table 4. Urban tree diversity classification averaged overall accuracies (AOA) and averaged Kappa index of agreement (AKIA) obtained using the 2-folds learning method and applied on each reduced hyperspectral image, and on the original (hyperspectral and Sentinel2) image datasets with RF and SVM according to the four sub-levels for families and species datasets.
19 Families
(n indiv. > 4)
12 Families
(n indiv. > 10)
27 Species
(n indiv. > 4)
14 Species
(n indiv. > 10)
Mean Values
AOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIA
SVMOriginal images36.40%± 3.60.1839.18%±4.10.2235.55%±3.50.1641.53%±3.00.1938.16%±3.60.19
PCA36.44%± 4.30.1838.86%±3.90.2034.54%±3.50.1440.04%±3.30.1737.47%±3.70.17
MNF54.17%± 6.00.4659.00%±5.90.5151.76%±4.90.4458.72%±4.30.4855.91%±5.30.47
VI ANOVA40.92%± 3.30.2845.15%±3.00.3337.63%±3.70.2344.37%±4.30.2942.02%±3.60.28
VI by PCA37.90%± 2.00.2441.34%±2.20.2736.14%±4.70.2042.54%±3.00.2739.48%±3.00.24
Mean values41.17%±3.80.2744.70%±3.80.3039.12%±4.00.2345.44%±3.60.2842.61%±3.80.27
AOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIA
RFOriginal images37.64%± 3.80.2540.99%±3.80.2835.76%±4.00.2341.74%±4.80.2639.03%±4.10.25
PCA37.21%± 4.60.2239.91%±4.80.2436.85%±3.30.2044.19%±4.50.2539.54%±4.30.23
MNF53.02%± 5.80.4457.43%±5.90.4851.36%±4.60.4258.54%±4.40.4855.09%±5.20.45AOA
VI ANOVA48.30%± 3.70.3952.21%±3.80.4245.08%±3.50.3653.01%±2.80.4349.65%±3.50.4050–60%
VI by PCA40.35%± 3.30.2942.13%±3.30.3040.98%±5.50.3046.37%±4.20.3442.46%±4.10.3160–70%
Mean values43.30%±4.20.3246.53%±4.30.3442.00%±4.20.3048.77%±4.10.3545.15%±4.20.33>70%
Table 5. Urban tree diversity classification results obtained using the 10-folds learning method and applied on reduced MNF image datasets, with SVM and RF classifiers according to subsets exhibiting at least 10 individuals per class.
Table 5. Urban tree diversity classification results obtained using the 10-folds learning method and applied on reduced MNF image datasets, with SVM and RF classifiers according to subsets exhibiting at least 10 individuals per class.
12 Families (n > 10)14 Species (n > 10)
Spatial ResolutionSVMRFSVMRFMean Values
OA σKIAOA σKIAOA σKIAOA σKIAOA σKIA
HySpex2 m73.81%±7.50.6873.81%±6.00.6878.38%±2.50.7375.68%±2.50.7075.42%±4.60.70
HYPXIM4 m73.81%±5.00.6873.81%±6.50.6778.38%±1.50.7372.97%±0.50.6574.74%±3.40.68AOA
8 m59.52%±4.00.4961.90%±3.00.5162.16%±3.00.5164.86%±3.50.5562.11%±3.40.5250–60%
Sentinel210 m61.90%±5.50.5359.52%±4.50.4959.46%±5.50.4862.16%±3.00.5260.76%±4.60.5160–70%
Mean values67.26%±5.50.6067.26%±5.00.5969.60%±3.10.6168.92%±2.40.6168.26%±4.00.60> 70%
Table 6. Urban tree diversity classification results obtained using the 2-folds learning method and applied on reduced MNF image datasets, with SVM and RF classifiers according to subsets exhibiting at least 10 individuals per class.
Table 6. Urban tree diversity classification results obtained using the 2-folds learning method and applied on reduced MNF image datasets, with SVM and RF classifiers according to subsets exhibiting at least 10 individuals per class.
12 Families (n > 10)14 Species (n > 10)
Spatial ResolutionSVMRFSVMRFMean Values
OA σKIAOA σKIAOA σKIAOA σKIAOA σKIA
HySpex2 m65.13%± 8.50.5864.08%± 8.00.5763.20%±5.50.5562.29%±5.00.5363.67%±6.80.56
HYPXIM4 m61.96%± 4.50.5560.71%± 6.50.5363.88%±4.00.5661.07%±3.00.5261.90%±4.50.54AOA
8 m59.21%± 5.50.5156.07%± 5.00.560.85%±4.00.558.27%±4.50.558.60%±4.80.4950–60%
Sentinel210 m49.70%± 5.00.3948.86%± 4.00.3846.95%±3.50.3052.56%±5.00.449.52%±4.40.3760–70%
Mean values59.00%±5.90.5157.43%±5.90.4858.72%±4.30.4858.54%±4.40.4858.42%±5.10.49> 70%
Table 7. Classifications mean results by spatial resolution and dimension reductions using the 2-folds learning method.
Table 7. Classifications mean results by spatial resolution and dimension reductions using the 2-folds learning method.
HySpex (2 m)HYPXIM (4 m)HYPXIM (8 m)Sentinel2 (10)Mean Values
AOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIAAOA σAKIA
SVMOriginal images51.67%±4.10.4230.29%±3.40.0430.94%±2.60.0339.76%±4.10.2738.16%±3.60.19
PCA49.95%±4.50.3932.74%±4.00.1231.63%±2.40.0438.05%±4.50.2538.09%±3.80.20
MNF60.79%±6.90.5359.29%±5.00.5256.85%±4.40.4947.96%±4.10.3756.22%±5.10.48
VI ANOVA46.62%±3.80.3645.00%±3.80.3434.44%±3.30.16 42.02%±3.60.28
VI by PCA38.60%±2.30.2142.10%±3.50.3140.20%±3.80.28 40.30%±3.20.27
RFOriginal images44.71%±3.30.3341.13%±4.90.2835.97%±4.10.2034.31%±4.00.2139.03%±4.10.25
PCA48.19%±5.30.3632.74%±4.00.1233.66%±3.30.1143.57%±4.60.3339.54%±4.30.23
MNF60.30%±6.80.5258.08%±4.90.4954.02%±4.90.4347.96%±4.10.3755.09%±5.20.45
VI ANOVA54.56%±4.50.4649.26%±3.00.3945.13%±2.90.34 49.65%±3.50.40AOA
VI by PCA41.54%±5.30.2944.61%±3.30.3441.21%±3.80.29 42.46%±4.10.3150–60%
Mean values (without VI)52.60%±5.10.4242.38%±4.40.2640.51%±3.60.2241.93%±4.30.3044.31%±4.10.3060–70%
Mean values49.69%±4.70.3943.52%±4.00.2940.40%±3.50.24 44.06%±4.00.31> 70%
Table 8. PA and UA for the MNF-reduced HySpex and HYPXIM 4 m hyperspectral images, classified with SVM and trained with the 14 species training dataset with 2-folds and 10-folds. These PA and UA correspond to those obtained for classifications shown in Figure 5 and Figure 6, exhibiting the greatest OA.
Table 8. PA and UA for the MNF-reduced HySpex and HYPXIM 4 m hyperspectral images, classified with SVM and trained with the 14 species training dataset with 2-folds and 10-folds. These PA and UA correspond to those obtained for classifications shown in Figure 5 and Figure 6, exhibiting the greatest OA.
2-Folds10-Folds
Number of IndividualsHySpex 2 mHYPXIM 4 mHySpex 2 mHYPXIM 4 m
PAUAPAUAPAUAPAUA
SVMMaple100.0%0.0%0.0%0.0%40.0%66.7%20.0%50.0%
Oak1216.7%33.3%41.7%71.4%58.3%87.5%41.7%62.5%
Fir137.7%25.0%0.0%0.0%38.5%35.7%23.1%20.0%
Shrub mix140.0%0.0%28.6%44.4%21.4%27.3%35.7%55.6%
Dwarf palm166.3%25.0%50.0%57.1%25.0%26.7%37.5%66.7%
Southern magnolia1625.0%33.3%62.5%55.6%56.3%52.9%50.0%53.3%
Lawn1747.1%80.0%52.9%69.2%58.8%71.4%58.8%62.5%
Silver linden170.0%0.0%0.0%0.0%23.5%40.0%5.9%50.0%
Horse chesnut2176.2%94.1%85.7%100.0%90.5%93.5%85.7%94.7%
Black locust2133.3%58.3%47.6%45.5%42.9%42.9%38.1%61.5%
Common privet2254.5%46.2%63.6%50.0%59.1%61.9%59.1%52.0%
Mediterranean hackberry3565.7%46.9%80.0%66.7%65.7%62.2%71.4%62.5%
Linden6387.3%49.1%55.6%43.8%69.8%57.9%74.6%46.5%
London plane14495.1%83.5%88.9%84.8%92.4%88.7%89.6%89.0%
OA63.2%63.9%78.4%78.4%

Share and Cite

MDPI and ACS Style

Brabant, C.; Alvarez-Vanhard, E.; Laribi, A.; Morin, G.; Thanh Nguyen, K.; Thomas, A.; Houet, T. Comparison of Hyperspectral Techniques for Urban Tree Diversity Classification. Remote Sens. 2019, 11, 1269. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11111269

AMA Style

Brabant C, Alvarez-Vanhard E, Laribi A, Morin G, Thanh Nguyen K, Thomas A, Houet T. Comparison of Hyperspectral Techniques for Urban Tree Diversity Classification. Remote Sensing. 2019; 11(11):1269. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11111269

Chicago/Turabian Style

Brabant, Charlotte, Emilien Alvarez-Vanhard, Achour Laribi, Gwénaël Morin, Kim Thanh Nguyen, Alban Thomas, and Thomas Houet. 2019. "Comparison of Hyperspectral Techniques for Urban Tree Diversity Classification" Remote Sensing 11, no. 11: 1269. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11111269

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop