Next Article in Journal
Optimal Scheduling and Fair Service Policy for STDMA in Underwater Networks with Acoustic Communications
Next Article in Special Issue
An Automatic and Novel SAR Image Registration Algorithm: A Case Study of the Chinese GF-3 Satellite
Previous Article in Journal
An Advanced IoT-based System for Intelligent Energy Management in Buildings
Previous Article in Special Issue
A Novel Ship Detection Method Based on Gradient and Integral Feature for Single-Polarization Synthetic Aperture Radar Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gaofen-3 PolSAR Image Classification via XGBoost and Polarimetric Spatial Information

1
School of Electronic Information, Wuhan University, Wuhan 430072, China
2
Collaborative Innovation Center of Geospatial Technology, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Submission received: 31 December 2017 / Revised: 13 February 2018 / Accepted: 14 February 2018 / Published: 17 February 2018
(This article belongs to the Special Issue First Experiences with Chinese Gaofen-3 SAR Sensor)

Abstract

:
The launch of the Chinese Gaofen-3 (GF-3) satellite will provide enough synthetic aperture radar (SAR) images with different imaging modes for land cover classification and other potential usages in the next few years. This paper aims to propose an efficient and practical classification framework for a GF-3 polarimetric SAR (PolSAR) image. The proposed classification framework consists of four simple parts including polarimetric feature extraction and stacking, the initial classification via XGBoost, superpixels generation by statistical region merging (SRM) based on Pauli RGB image, and a post-processing step to determine the label of a superpixel by modified majority voting. Fast initial classification via XGBoost and the incorporation of spatial information via a post-processing step through superpixel-based modified majority voting would potentially make the method efficient in practical use. Preliminary experimental results on real GF-3 PolSAR images and the AIRSAR Flevoland data set validate the efficacy and efficiency of the proposed classification framework. The results demonstrate that the quality of GF-3 PolSAR data is adequate enough for classification purpose. The results also show that the incorporation of spatial information is important for overall performance improvement.

1. Introduction

The Chinese Gaofen-3 (GF-3) satellite carrying a synthetic aperture radar (SAR) sensor, was launched on 10 August 2016. With a powerful imaging capability of 12 modes, GF-3 can provide polarimetric SAR (PolSAR) images with a nominal resolution of 8 m. By transmitting and receiving horizontal and vertical waves, PolSAR data is able to capture four polarization channels, helpful for physical understanding and mechanism interpretation regarding land surface [1]. Many applications such as crop monitoring, biomass estimation, target detection, land cover and land use classification, can be undertaken by choosing appropriate GF-3 SAR data. Among the applications, the classification of PolSAR image plays an important role in PolSAR image analysis. Actually, land cover classification on a real whole PolSAR image is the main task in our cooperation with the National Disaster Reduction Center of China (NDRCC). How to find an efficient framework to classify an entire PolSAR image quickly and accurately is the main concern in practical use.
In the last two decades, researchers have developed many PolSAR image classification algorithms. In general, these algorithms can be divided into three categories called statistical model-based algorithms [2,3,4], scattering mechanism analysis-based algorithms [5,6,7], and algorithms combining standard classifiers and polarimetric features [8]. Many works focus on the third type among these approaches. More and more complicated and efficient classifiers or features are integrated into the PolSAR image classification framework to improve accuracy [9,10,11,12]. For example, many neural network(NN)-based methods are proposed and they have reported very excellent results in recent years [11,13,14,15]. However, even though accuracy is improved, most classification algorithms and procedures of complicated feature extraction are of high complexity and time-consuming, which makes them impractical to process an entire PolSAR image. Moreover, enough labeled samples for NN-based classifiers are not always available. Therefore, an efficient classifier with high accuracy and low time cost is preferable in practical use, which also explains why tree-based classifiers are still welcome in remotely sensed image analysis [16,17,18,19,20]. In Ref. [16], QUEST (quick, unbiased, efficient, statistical tree [21]) was used as a decision tree (DT) tool to implement the classification. The decision tree was very efficient and provided clear split rules that can be easily interpreted according to the physical understanding of used features. Du et al. [17] used random forest (RF) and rotation forest to investigate polarimetric-spatial features in PolSAR image classification and found that rotation forest performed better while RF was much faster [17]. Pradhan et al. [18] compared the performance of several classification techniques for extracting urban areas. It turned out that the rule-based classifier and DT distinguished land cover classes better than support vector machine (SVM) and K-nearest neighbor (KNN) [18]. In [19,20], hybrid DT and RF were used as classifiers, separately, and showed comparable results. Previous work has shown that the tree-based classifier is very efficient and popular in practical use.
Polarimetric features are often associated with the scattering mechanism of terrain scatterers and represent the information embedded in PolSAR data. The quality of input features basically determines the classification performance of employed classifiers. Besides polarimetric information, spatial information is another important factor that affects classification performance and has been paid more and more attentions. The introduction of spatial features in PolSAR image analysis is inspired by the remarkable improvements resulting from the complementarity between spectral and spatial features in optical image classification [17]. Markov random fields (MRF) modeling, texture information and region-based or object-based algorithms are the three main methodological approaches that can be employed to incorporate spatial information into image analysis [22,23]. Du et al. combined spatial features (consisting of textural metrics and morphological profiles) with polarimetric features for PolSAR image classification [17]. In our previous work, spatial information and statistical model-based data fidelity were incorporated into the classification framework via MRF [24]. In Ref. [25], superpixels were generated by a modified linear iterative clustering algorithm to incorporate spatial context. Another potential benefit due to the incorporation of spatial information is to reduce speckles, which are very common in PolSAR images and adversely affect the classification performance.
To make the classification method more practical in use and incorporate spatial information at each image pixel, we select a recently proposed tree-based classification framework called XGBoost [26] for initial image classification. Spatial information is incorporated via superpixels generated by the statistical region merging (SRM) method[27]. After superpixels are generated, the label of a superpixel is determined in a post-processing step through a modified majority voting strategy, which runs quickly and turns out to be efficient. The remainder of this paper is organized as follows. In Section 2, we introduce the adopted features, related classifiers and the proposed classification framework. In Section 3, we describe the experimental data sets and analyze the classification results obtained on real GF-3 PolSAR images. Additional discussions are presented in Section 4. Conclusions are drawn in Section 5.

2. Method

2.1. Polarimetric Features

Providing more information for land cover classification, PolSAR data can be represented by a scattering matrix S, a covariance matrix C or a coherency matrix T [1]. | H H | 2 , | H V | 2 , and | V V | 2 represent the scattered power of three channels and directly reflect the type of scattering media. With an assumption of reciprocal condition S H V = S V H , the coherency matrix can be expressed by T 3 , of which the diagonal elements are T 11 , T 22 , and T 33 , respectively.
T 11 = 0.5 | S H H + S V V | 2 ,   T 22 = 0.5 | S H H S V V | 2 ,   T 33 = 2 | S H V | 2
Another useful polarimetric parameter is the circular polarization correlation coefficient ρ [28], which can be expressed as
ρ = < S R R S L L > < | S R R | 2 > < | S L L | 2 >
Man-made targets such as buildings exhibit low surface roughness and a high value of | ρ | [29]. Many other techniques also have been proposed to decompose PolSAR data and most of these can be divided into two categories [30]. One category is based on an eigenvector-based decomposition proposed by Cloude [5]. Three eigenvector-based parameters of entropy H, scattering angle α , and anisotropy A can classify a pixel by scattering mechanisms.
H = i = 1 3 p i log 3 p i , α = i = 1 3 p i α i , A = λ 2 λ 3 λ 2 + λ 3
where p i correspond to the pseudo-probabilities from the eigenvalues, λ i are the sorted eigenvalues of T 3 , and α i represent scattering mechanism associated with each eigenvector. The other category is model-based decomposition such as Freeman decomposition, which models PolSAR data as the sum of surface, volume, and double-bounce scattering components under an assumption of reflection symmetry [6]. Yamaguchi et al. went further than the Freeman decomposition by introducing the helix component [7].
s p a n = Y P S + Y P V + Y P D + Y P C
However, it has been demonstrated that the Yamaguchi decomposition may generate negative powers due to overestimation in the volume scattering, which is obviously nonphysical [30]. Van Zyl et al. tackled the problem by introducing the nonnegative eigenvalue decomposition (NNED) and ensured nonnegative scattering powers of different scattering mechanisms [30].
In this paper, we extract 16 polarimetric features to represent the information embedded in GF-3 PolSAR data. Due to the relation between T 33 and | H V | 2 , only T 33 is taken into consideration. Even though Van Zyl NNED and Yamaguchi decompositions reflect similar scattering mechanisms, we simultaneously use the two decompositions since the extraction procedures are different and Yamaguchi decomposition provides another helix component. Therefore, the total selected features consist of | H H | 2 , | V V | 2 , the diagonal elements of T 3 ( T 11 , T 22 , T 33 ), | ρ | , three eigenvector-based parameters from H / α / A model (H, α , A), four components of Yamaguchi decomposition ( Y P S , Y P V , Y P D , Y P C ) and three components of Van Zyl NNED decomposition ( V P S , V P V , V P D ).

2.2. XGBoost for PolSAR Image Classification

In PolSAR image classification, the goal is to determine the labels of unknown scattering cells or pixels based on a fine-trained classifier. The adopted XGBoost is short for “extreme gradient boosting” and designed to be a scalable machine learning system for tree boosting. The system runs much faster and gives state-of-the-art results on many problems from data mining challenges [26]. Given n labeled samples with m features D = ( x i , y i ) ( | D | = n , x i m , y i ) , the tree ensemble method uses K additive functions to predict the label.
y ^ i = k = 1 K f k ( x i ) , f k F
where F = { f ( x ) = w q ( x ) } ( q : m T , w T ) is the space of regression trees. q represents the structure of each tree that has T leaves. Each f k corresponds to an independent tree structure q and leaf weights w. To learn the set of functions in the model, XGBoost minimizes the following regularized objective.
L = i l ( y ^ i , y i ) + k Ω ( f k ) , w h e r e Ω ( f ) = γ T + 1 2 λ | | w | | 2
Here l is the loss function and Ω is the regularized term. The regularized objective is slightly improved compared to previous gradient tree boosting algorithm. Let y ^ i ( t ) be the prediction of the i-th instance at the t-th iteration. The ensemble model works better in an additive manner. Second-order approximation is used to speed up the optimization procedure.
L ( t ) = i = 1 n l ( y i , y ^ i ( t 1 ) + f t ( x i ) ) + Ω ( f t )
L t i = 1 n [ l ( y i , y ^ i ( t 1 ) ) + g i f t ( x i ) + 1 2 h i f t 2 ( x i ) ] + γ T + 1 2 λ j = 1 T w j 2
where g i and h i are first and second order gradient statistics on the loss function. For a fixed structure q ( x ) , the optimal weight w and the corresponding optimal value can be calculated by evaluating the split candidates.
Besides the improvements in the regularized objective, several additional techniques are also used to further promote classification performance. Shrinkage and column/feature subsampling are used to prevent overfitting. Column/feature subsampling can also speed up computations of the algorithm. In order to find the best split efficiently, the algorithm visits feature values in sorted order to accumulate the gradient statistics in (8) and an approximate algorithm is summarized to avoid enumerating all possible splits greedily. In system design level, a block structure is used to reduce the cost of sorting, which is the most time consuming part of tree learning. Readers can refer to [26] for detailed information of XGBoost. In case of the mentioned advancements and excellent performance in practical use, XGBoost is adopted for PolSAR image to quickly generate the initial classification map.

2.3. Post-Processing via Superpixels and Majority Voting

To incorporate spatial information into the pixel-wise classification framework, we adopt the decision strategy of superpixel-based majority voting, which is similar to [31] except for the decision rule and the generation algorithm of superpixels. Superpixels can be generated via segmentation, which is another important application of PolSAR images. Many superpixel segmentation algorithms suitable for PolSAR data have been proposed by incorporating polarimetric information. Lang et al. improved SRM considering the characteristics of PolSAR data and made the generalized SRM suitable for single- or multi-dimensional SAR data [32]. Xiang et al. modified the simple linear iterative clustering(SLIC) and developed an adaptive superpixel generation procedure with local iterative clustering and spherically invariant random vector (SIRV) [33]. Wang et al. introduced two distance measures and an entropy rate method into the PolSAR image superpixel segmentation [34]. In this paper, statistical region merging is directly applied to the Pauli RGB image to generate superpixels for its rapidity [27,32]. The parameter Q in SRM quantifies the statistical complexity of a perfect scene I and controls the coarseness of the segmentation. Conceptual simplicity and the ability of coping with significant noise corruption make SRM suitable for superpixel generation of PolSAR image. One way to use superpixels is to apply classification algorithms on superpixels directly. But it is inconvenient for users to select superpixels as training samples and may also decrease the number of available samples. In this paper, the post-processing procedure of superpixel-based majority voting is shown in Figure 1. The direct use of superpixel-based majority voting goes both ways. It may misclassify a superpixel if the majority is misclassified. To weaken such an effect, we modify the majority voting. For a given superpixel s p j with n pixels that are classified into C classes, we construct the normalized histogram H = { H 1 , H 2 , , H C } of class labels and the sorted histogram Hs = { H s 1 , H s 2 , , H s C } in descending order. If H s 1 < = σ 1 and H s 2 > = σ 2 , the pixels labeled s 2 are kept and the remaining pixels are assigned as the majority label. Otherwise, the label with the highest frequency is assigned to all pixels within the superpixel s p j .
C s p j = arg   max i { 1 , 2 , , C } ( H i )
The proposed classification framework consists of four simple parts including polarimetric feature extraction and stacking, the initial classification via XGBoost, superpixels generation by SRM based on the Pauli RGB image, and label determination via a post-processing step through superpixel-based modified majority voting. The extended Lee Sigma filter is actually used to pre-process the original GF-3 PolSAR images for speckle filtering [35,36]. The workflow of the method is presented in Figure 2. Fast initial classification via XGBoost and the incorporation of spatial information in the post-processing step via superpixel-based modified majority voting, would potentially make the method efficient in practical use and promote the classification performance.

3. Experiments and Results Analysis

3.1. Experimental Data

In this study, the city of Wuhan, which lies in central China, is chosen as a case study area. Two GF-3 single-look complex (SLC) PolSAR images with quad-polarized strip I (QPSI) mode are captured to evaluate the proposed algorithm. Both images are acquired with similar imaging parameters over Wuhan, China, as shown in Figure 3. The detailed data information is presented in Table 1. Two sub-images specified by the red rectangles in Figure 3 are cropped from the two whole images (images #1 and #2) and we name the sub-images data set A and data set B, respectively. Two optical images acquired by Sentinel-2 [37] on 26 March 2017 are also collected for identifying land cover classes. Figure 4 shows the mosaic image and the two regions of interest.
Figure 5 illustrates Pauli RGB images of the two data sets. Data set A has a size of 656× 1192 pixels and covers the whole region of Jinyin Lake, Wuhan. Three classes consisting of water, buildings, and vegetation, are identified, as shown in Figure 5b. Data set B has a size of 603 × 976 pixels. The study site is located at a rural region of Hannan District, Wuhan. A total of five classes are considered, which include water, building, farmland, vegetation1 and vegetation2. Vegetation areas are divided into two classes according to the backscattering responses and vegetation types, which is shown in Figure 5d. Since only Sentinel-2 optical images are available and no field survey is conducted, the exact land cover types are not identified. It is speculated that vegetation1 are shrubberies or trees with strong branches and vegetation2 involve grass or reeds that grow in wet places. Vegetation1 show stronger scattered power than vegetation2. Both ground-truth maps shown in Figure 5 are manually created based on the Sentinel-2 optical images in Figure 3. The white areas labeled “None” in both data sets are not included in the final accuracy assessment step.

3.2. Classification Results

In this section, initial classification results and related analysis achieved by the proposed framework from real GF-3 PolSAR data are reported. The compared classifiers include DT, RF, and SVM with radial basis function kernel. The parameters of three tree-based classifiers are set default for fair comparison. Parameter tuning is recommended for tree-based classifiers to produce better classification results. A total of n = 2000 labeled samples are randomly selected to train the classifiers. The number of train samples for each class is equal. Remaining samples are used for evaluation purpose. The extended Lee sigma filter with σ v = 0.9 and a 9 × 9 window is adopted for speckle reduction of GF-3 PolSAR data [36]. The empirical parameters of modified majority voting are set σ 1 = 0.8 and σ 2 = 0.1 . m = 16 polarimetric features are employed as described in Section 2.1. Overall accuracy (OA) and the confusion matrix are adopted to evaluate the performance.
Samples of five classes consisting of water, building, farmland, vegetation1 and vegetation2, are randomly selected from data set B, image #2. Figure 6 illustrates the 2D scatter plots of different features in pairs. All the features are extracted by PolSARPro 5.1. Water and farmland show higher surface components while the volume components of both vegetation types are obvious. Figure 6g illustrates the scatter plots of five classes basically fit into the zones defined in the H / α plane [5]. However, it also shows that the used GF-3 PolSAR image shows higher values of the entropy H. The | ρ | value, the anisotropy A and the double-bounce component of buildings are higher compared to other land cover classes. The components of Yamaguchi and Van Zyl NNED decompositions reflect similar scattering mechanisms according to corresponding scatter plots. The characteristics of adopted features from GF-3 PolSAR data are basically in accordance with the theorems.
Figure 7 and Figure 8 illustrate the pixel-wise classification maps on the two data sets, respectively. The OAs obtained from two data sets are shown in Table 2 and Table 3. According to the results, The OAs of DT on both data sets are worse. The other three classifiers preform well and report comparable results on both data sets. For both data sets, XGBoost performs best. From Table 3, it can be seen that the improperly classified regions mainly occur in the building area. The corresponding accuracies obtained by the four classifiers are all below 77%. Figure 9 shows the confusion matrices of both data sets, respectively. For data set A, the misclassification mainly occurs between building and vegetation types. For data set B, about 20% of the building area are misclassified as vegetation areas. The reason is that data set B covers rural region, where trees are planted around buildings. Misclassification between farmland, vegetation1 and vegetation2 results from their similarity. In all, XGBoost reports similar classification results to SVM and RF, but performs much better than DT. However, many misclassified pixels emerge in homogeneous areas for all classification maps, which badly affect the performance. Moreover, the classification results in building areas are not adequate. Some necessary steps should be taken to remove the discrete misclassified pixels and to improve the the overall performance.

3.3. Performance Analysis

To further compare performances of different classifiers, more experiments are conducted here. First, the effects of the number of training samples are examined. Figure 10 illustrates the overall classification accuracies for each classifier under different numbers of training samples. 10 to 2000 pixels of the labeled data are randomly selected as training samples. Additionally, 5000 randomly selected pixels are used for testing. Ten repeated experiments are conducted and we take the average as the final OA. Classification accuracies of both data sets generally increase as more training samples are selected. Other conclusions are same to Section 3.2. XGBoost, SVM, and RF report similar results and perform much better than DT.
Figure 11 shows the time costs of different classifiers. Both training samples and test samples are randomly selected. The training sample numbers is set to 2000. The number of test samples ranges from 1000 to 10,000. Accordingly, The curves of time costs on two data sets are very similar with each other. The time cost of SVM linearly increases as more samples are used. The time cost of DT is lowest but the classification result is not satisfying. SVM costs more time than XGBoost but report similar classification results. RF costs more time than XGBoost on data set A while the contrary is the case on data set B. XGBoost performs better than RF on both data sets according to Figure 10. In sum, XGBoost is most efficient in case of classification result and time cost.

3.4. Final Classification

Based on the initial classification results obtained by XGBoost, the post-processing step is conducted to incorporate spatial information into the final classification map via superpixels and the modified majority voting. Superpixels are generated by SRM based on the Pauli RGB image. The parameter Q in SRM controls the coarseness of superpixels. The post-processing step can also be combined with other classifiers. Figure 12 shows the change of OA for different classifiers when the Q level increases from 1 to 256. Table 4 and Table 5 report the Q levels with best OAs on the two data sets, respectively. It is visible that the OAs for both data sets have improved, which demonstrates the effectiveness of incorporating spatial information through superpixels for pixel-wise classification. It can also concluded that the Q level has an impact on the classification result. One point is the Q level should not be set too small.The Q levels with highest OAs for the two data sets are different and should be set accordingly.
Figure 13 and Figure 14 show the generated superpixels and the final classification maps of both data sets with selected Q levels, according to the OA curves in Figure 12. The specific Q levels can be found at Table 4 and Table 5. Compared to Figure 7 and Figure 8, the misclassified pixels in a superpixel have been removed, which helps to generate smoothed classification maps. The confusion matrices shown in Figure 15 illustrate detailed changes after the incorporation of spatial information. For data set A, the percentage at which buildings are wrongly assigned as vegetations decreases by 19.42%. The accuracy of building reaches 96.38%. The OA of data set A acquired by XGBoost is 92.20%, which is superior to other classifiers. As for data set B, the accuracies of building, farmland, and vegetations increased. Specifically, the accuracy of farmland areas in data set B is higher by 11% than that of the pixel-wise results. Overall performance and accuracies of most classes improve at the price of the misclassification between building and farmland areas. The incorporation of spatial information significantly contributes to the performance improvement. Even though SVM and XGBoost report comparable OAs according to Table 5, tree-based classifiers, especially XGBoost, are preferable in practical classification application in consideration of accuracy and time costs.

4. Discussions

4.1. Discussion on Polarimetric Features

In this section, the two GF-3 PolSAR data sets are further processed with six feature sets (FS) to find out how these features contribute to classification results. The detailed content of feature sets can be seen in Table 6. In order not to be affected by spatial information, initial classification results without the post-processing step are analyzed. Table 7 and Table 8 report the accuracies by classes and OAs with the six feature sets on both data sets. FS1 produces the lowest OAs and the lowest accuracies of building on both data sets. The H / α / A model improves the results obviously by increasing OAs by 3.96% and 5.4%, respectively. | ρ | contributes to building areas but has not benefited much to overall classification results. Both Yamaguchi and Van Zyl NNED decompositions contribute to building areas and further promote the overall performance. However, the simultaneous use of Yamaguchi and Van Zyl NNED decompositions decreases the accuracy of buildings. In all, H, α , and A are three important parameters for land cover classification. One of the Yamaguchi, Van Zyl NNED or other model-based decompositions reflecting similar scattering mechanisms is recommended.

4.2. Discussion on Road Type

In data set A, the roads surrounding Jinyin Lake often appear as vegetation types. This is probably due to the lack of training samples since current land cover classes do not contain roads. Another possible reason is that trees are planted along the roads as shown in Figure 16. Here, some training sites on the road around Jinyin Lake are added into the original ground truth map. Figure 17 shows the updated ground truth map and final classification map with the road type. Some parts of roads are extracted. Some none-road areas are unexpectedly misclassified as roads. Based on the confusion matrix shown in Figure 18, the accuracy of road is only 32.70% and the accuracies acquired by the other three compared classifiers are 29.20% (DT), 34.56% (RF), and 21.21% (SVM). The majority of misclassification can be found in vegetation and building areas for trees and buildings alongside roads. The classification methods are unsuitable for the extraction of roads or other line targets.

4.3. Discussion on More Complex Land Cover Classification System

To further evaluate the performance on the data set with a more complex land cover classification system, the AIRSAR Flovland data set is tested by the proposed classification framework. The scene covers over Flevoland, the Netherland with an image size of 750 × 1024 pixels. With well-established ground truth map, the AIRSAR Flevoland data set has been widely used in land use classification since [38]. The land cover classification system consists of eight crop classes and three other classes of bare soil, water, and forest, which is shown in Figure 19. In consideration of the homogeneity of this data set, the empirical parameters of the modified majority voting are set σ 1 = 0.5 and σ 2 = 0.4 . The extended Lee sigma filter is adopted and other parameter settings are same to Section 3.
Figure 20 illustrates the change of OA with different Q levels. Table 9 reports the Q level with best OA for different classifiers. The OAs acquired by methods proposed by Lee et al. [38], Tao et al. [10], and Qin et al. [11] are also reported in Table 9 since the same data set has been used and land cover classes are same. Lee et al. used an maximum likelihood classifier based on the Wishart distribution for all polarization combinations to quantitatively assess classification results [38]. Tao et al. developed a classification method with selected hidden polarimetric features in the rotation domain and a SVM/DT classifier. The overall classification accuracy of DT with a window size of 9 × 9 is shown in Table 9 in view of better performance and fair comparison [10]. Qin et al. developed an object-oriented classification method using restricted Boltzmann machines. The best OA of the proposed RBM-Adaboost with 29 units and 88 learners is 96.15% [11]. According to the results, tree-based classifiers surpass SVM by 3%. The OA of XGBoost is higher by 10% than that of Lee’s method. Figure 21 shows the generated superpixels and the final classification map acquired by XGBoost. The corresponding confusion matrix is shown in Figure 22. The accuracies of most classes except for beet, are above 90%. Therefore, the proposed method is competitive with other benchmark methods.

5. Conclusions

In this paper, we proposed an efficient classification framework for GF-3 PolSAR images. With quick initial classification and incorporation of spatial information, the proposed method is hopefully able to classify an entire GF-3 PolSAR image quickly and accurately. We also analyzed the characteristics of features from GF-3 PolSAR data and how the adopted features contribute to classification results. Preliminary experimental results on real GF-3 PolSAR images and the AIRSAR Flevoland PolSAR data have demonstrated their own advantages. XGBoost showed comparable classification results with less time consumed. Spatial information turned out to be an important factor for PolSAR image classification and accounted for the accuracy improvements achieved in the post-processing step via superpixel-based modified majority voting. The results also showed that the quality of GF-3 PolSAR data was adequate enough for land cover classification and other potential usages.
Further improvement is still required. Analysis about more palarimetric parameters extracted from Gaofen-3 data [39] and how to choose essential polarimetric features need to be investigated. The generation of adequate superpixels is worth further exploration. Future efforts should also be devoted to implementing the classification framework in C++, which will further improve efficiency.

Acknowledgments

This work was supported by the Technology Research and Development of the Major Project of High-Resolution Earth Observation System under Grant 03-Y20A10-9001-15/16. It was also supported by the Fundamental Research Funds for the Central Universities under Grant NO.2042017kf0211. The authors thank the anonymous reviewers for their helpful comments and suggestions. We also appreciate China Center For Resources Satellite Data and Application for providing GF-3 PolSAR images.

Author Contributions

All the authors made significant contributions to the work. Hao Dong conducted the experiments and wrote the paper. Xin Xu helped design the experiments and the paper structure. Lei Wang collected the data and analyzed the experimental results. Xin Xu and Fangling Pu provided suggestions for the preparation and revision of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, J.S.; Pottier, E. Overview of polarimetric radar imaging. In Polarimetric Radar Imaging: From Basics to Applications; CRC Press: Boca Raton, FL, USA, 2009; pp. 1–4. ISBN 978-1-4200-5497-2. [Google Scholar]
  2. Lee, J.S.; Grunes, M.R.; Kwok, R. Classification of multi-look polarimetric SAR imagery based on complex Wishart distribution. Int. J. Remote Sens. 1994, 15, 2299–2311. [Google Scholar] [CrossRef]
  3. Freitas, C.C.; Frery, A.C.; Correia, A.H. The polarimetric G distribution for SAR data analysis. Environmetrics 2005, 16, 13–31. [Google Scholar] [CrossRef]
  4. Gao, W.; Yang, J.; Ma, W. Land cover classification for polarimetric SAR images based on mixture models. Remote Sens. 2014, 6, 3770–3790. [Google Scholar] [CrossRef]
  5. Cloude, S.R.; Pottier, E. An entropy based classification scheme for land applications of polarimetric SAR. IEEE Trans. Geosci. Remote Sens. 1997, 35, 68–78. [Google Scholar] [CrossRef]
  6. Freeman, A.; Durden, S.L. A three-component scattering model to describe polarimetric SAR data. IEEE Trans. Geosci. Remote Sens. 1998, 36, 963–973. [Google Scholar] [CrossRef]
  7. Yamaguchi, Y.; Moriyama, T.; Ishido, M.; Yamada, H. Four-component scattering for polarimetric SAR image decomposition. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1699–1706. [Google Scholar] [CrossRef]
  8. Gou, S.; Qiao, X.; Zhang, X.; Wang, W.; Du, F. Eigenvalue analysis-based approach for PolSAR image classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 805–818. [Google Scholar] [CrossRef]
  9. Uhlmann, S.; Kiranyaz, S. Integrating color features in polarimetric SAR image classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2197–2216. [Google Scholar] [CrossRef]
  10. Tao, C.; Chen, S.; Li, Y.; Xiao, S. PolSAR land cover classification based on roll-invariant and selected hidden polarimetric features in the rotation domain. Remote Sens. 2017, 9, 660. [Google Scholar]
  11. Qin, F.; Guo, J.; Sun, W. Object-oriented ensemble classification for polarimetric SAR Imagery using restricted Boltzmann machines. Remote Sens. Lett. 2017, 8, 204–213. [Google Scholar] [CrossRef]
  12. Wang, W.; Xiang, D.; Zhang, J.; Wan, J. Integrating contextual information with H/ α ¯ decomposition for PolSAR data classification. IEEE Geosci. Remote Sens. Lett. 2016, 13, 2034–2038. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Wu, L.; Neggaz, N.; Wang, S.; Wei, G. Remote-sensing image classification based on an improved probabilistic neural network. Sensors 2009, 9, 7516–7539. [Google Scholar] [CrossRef] [PubMed]
  14. Zhou, Y.; Wang, H.; Xu, F.; Jin, Y.Q. Polarimetric SAR image classification using deep convolutional neural networks. IEEE Trans. Geosci. Remote Sens. Lett. 2016, 13, 1835–1929. [Google Scholar] [CrossRef]
  15. Hou, B.; Kou, H.; Jiao, L. Classification of polarimetric SAR images using multilayer autoencoders and superpixels. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 3072–3081. [Google Scholar] [CrossRef]
  16. Qi, Z.; Yeh, A.G.; Li, X.; Lin, Z. A novel algorithm for land use and land cover classification using RADARSAT-2 polarimetric SAR data. Remote Sens. Environ. 2012, 118, 21–39. [Google Scholar] [CrossRef]
  17. Du, P.; Samat, A.; Waske, B.; Liu, S.; Li, Z. Random forest and rotation forest for fully polarized SAR image classification using polarimetric and spatial features. ISPRS J. Photogramm. Remote Sens. 2015, 105, 38–53. [Google Scholar] [CrossRef]
  18. Pradhan, B.; Abdullahi, S.; Seddighi, Y. Detection of urban environments using advanced land observing satellite phased array type L-band synthetic aperture radar data through different classification techniques. J. Appl. Remote Sens. 2016, 10, 036029. [Google Scholar] [CrossRef]
  19. Yang, G.; Fang, S. Improving remote sensing image classification by exploiting adaptive features and hierarchical hybrid decision trees. Remote Sens. Lett. 2017, 8, 156–164. [Google Scholar] [CrossRef]
  20. Mahdianpari, M.; Salehi, B.; Mohammadimanesh, F.; Motagh, M. Random forest wetland classification using ALOS-2 L-band, RADARSAT-2 C-band, and TerraSAR-X imagery. ISPRS J. Photogramm. Remote Sens. 2017, 130, 13–31. [Google Scholar] [CrossRef]
  21. Loh, W.Y.; Shih, Y.S. Split selection methods for classification trees. Statistica Sinica 1997, 7, 815–840. [Google Scholar]
  22. Ghamisi, P.; Mura, M.D.; Benediktssson, J.A. A survey on spectral-spatial classification techniques based on attribute profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  23. Moser, G.; Serpico, S.B.; Benendiktsson, J.A. Land cover mapping by Markov modeling of spatial-contextual information in very-high-resolution remote sensing images. Proc. IEEE 2013, 101, 631–651. [Google Scholar] [CrossRef]
  24. Dong, H.; Xu, X.; Sui, H.; Xu, F.; Liu, J. Copula-based joint statistical model for polarimetric features and its application in PolSAR image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5777–5789. [Google Scholar] [CrossRef]
  25. Xu, Q.; Chen, Q.; Yang, S.; Liu, X. Superpixel-based classification using K distribution and spatial context for polarimetric SAR images. Remote Sens. 2016, 8, 619. [Google Scholar] [CrossRef]
  26. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. arXiv, 2016; arXiv:1603.02754. [Google Scholar]
  27. Nock, R.; Nielsen, F. Statistical region merging. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1452–1458. [Google Scholar] [CrossRef] [PubMed]
  28. Mattia, F.; Le Toan, L.; Souyris, J.C.; Carolis, G.D.; Floury, N.; Posa, F. The effact of surface roughness on multifrequency polarimetric SAR data. IEEE Trans. Geosci. Remote Sens. 1997, 35, 954–966. [Google Scholar] [CrossRef]
  29. Li, X.; Guo, H.; Zhang, L.; Chen, X.; Liang, L. A new approach to collapsed building extraction using RadarSAT-2 polarimetric SAR imagery. IEEE Geosci. Remote Sens. Lett. 2012, 9, 677–681. [Google Scholar]
  30. Van Zyl, J.J.; Arii, M.; Kim, Y. Model-based decomposition of polarimetric SAR covariance matrices constrained for nonnegative eigenvalues. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3452–3459. [Google Scholar] [CrossRef]
  31. Feng, J.; Cao, Z.; Pi, Y. Polarimetric contextual classification of PolSAR images using sparse representation and superpixels. Remote Sens. 2014, 6, 7158–7181. [Google Scholar] [CrossRef]
  32. Lang, F.; Yang, J.; Li, D.; Zhao, L.; Shi, L. Polarimetric SAR image segmentation using statistical region merging. IEEE Geosci. Remote Sens. Lett. 2014, 11, 509–513. [Google Scholar] [CrossRef]
  33. Xiang, D.; Ban, Y.; Wang, W.; Su, Y. Adaptive superpixel generation for polarimetric SAR images with local iterative clustering and SIRV model. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3115–3131. [Google Scholar] [CrossRef]
  34. Wang, W.; Xiang, D.; Ban, Y.; Zhang, J.; Wan, J. Superpixel segmentation of polarimetric SAR images based on integrated distance measure and entropy rate method. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 4045–4058. [Google Scholar] [CrossRef]
  35. Lee, J.S.; Wen, J.H.; Ainsworth, T.L.; Chen, K.S.; Chen, A.J. Improved sigma filter for speckle filtering of SAR imagery. IEEE Trans. Geosci. Remote Sens. 2009, 47, 202–213. [Google Scholar]
  36. Lee, J.S.; Ainsworth, T.L.; Wang, Y.; Chen, K.S. Polarimetric SAR speckle filtering and the extended sigmma filter. IEEE Trans. Geosci. Remote Sens. 2009, 53, 1150–1160. [Google Scholar] [CrossRef]
  37. Open Access to Sentinel-1, Sentinel-2, and Sentinel-3 User Product. Available online: https://scihub.copernicus.eu (accessed on 14 September 2017).
  38. Lee, J.S.; Grunes, M.R.; Pottier, E. Quantitative comparison of classification capability: Fully polarimetric versus dual and single-polarization SAR. IEEE Trans. Geosci. Remote Sens. 2001, 39, 2343–2351. [Google Scholar]
  39. Xiang, D.; Tang, T.; Ban, Y.; Su, Y.; Kuang, G. Unsupervised polarimetric SAR urban area classification based on model-based decomposition with cross scattering. ISPRS J. Photogram. Remote Sens. 2016, 116, 86–100. [Google Scholar] [CrossRef]
Figure 1. Incorporation of spatial information in the post-posting step via superpixels and modified majority voting. Class types are represented by corresponding colors.
Figure 1. Incorporation of spatial information in the post-posting step via superpixels and modified majority voting. Class types are represented by corresponding colors.
Sensors 18 00611 g001
Figure 2. Workflow of the proposed method.
Figure 2. Workflow of the proposed method.
Sensors 18 00611 g002
Figure 3. Geographic extent of the study area and Pauli RGB images of the GF-3 PolSAR data (left: image #1, right: image #2). The red rectangles indicate the locations of the sub-images that are cropped.
Figure 3. Geographic extent of the study area and Pauli RGB images of the GF-3 PolSAR data (left: image #1, right: image #2). The red rectangles indicate the locations of the sub-images that are cropped.
Sensors 18 00611 g003
Figure 4. Optical image of the study area. Above is the mosaic Sentinel-2 optical image and the ROIs are masked by blue and red rectangles, respectively. On the bottom are the zoomed-in images of ROIs.
Figure 4. Optical image of the study area. Above is the mosaic Sentinel-2 optical image and the ROIs are masked by blue and red rectangles, respectively. On the bottom are the zoomed-in images of ROIs.
Sensors 18 00611 g004
Figure 5. Sub-images of the Gaofen-3 PolSAR Data. (a) Pauli RGB image of dataset A. (b) The ground-truth map of (a). (c) Pauli RGB image of dataset B. (d) The ground-truth map of (c).
Figure 5. Sub-images of the Gaofen-3 PolSAR Data. (a) Pauli RGB image of dataset A. (b) The ground-truth map of (a). (c) Pauli RGB image of dataset B. (d) The ground-truth map of (c).
Sensors 18 00611 g005
Figure 6. 2D scatter plots of samples corresponding to different polarimetric features. Each color represents one land cover type. (a) T 11 T 22 , (b) T 11 T 33 , (c) T 22 T 33 , (d) | H H | 2 | V V | 2 , (e) | H H | 2 | ρ | , (f) | V V | 2 | ρ | , (g) H α , (h) HA, (i) α A, (j) Y P S Y P V , (k) Y P S Y P D , (l) Y P V Y P D , (m) V P S V P V , (n) V P S V P D , (o) V P V V P D . The power-related features are divided by s p a n for visualization.
Figure 6. 2D scatter plots of samples corresponding to different polarimetric features. Each color represents one land cover type. (a) T 11 T 22 , (b) T 11 T 33 , (c) T 22 T 33 , (d) | H H | 2 | V V | 2 , (e) | H H | 2 | ρ | , (f) | V V | 2 | ρ | , (g) H α , (h) HA, (i) α A, (j) Y P S Y P V , (k) Y P S Y P D , (l) Y P V Y P D , (m) V P S V P V , (n) V P S V P D , (o) V P V V P D . The power-related features are divided by s p a n for visualization.
Sensors 18 00611 g006aSensors 18 00611 g006b
Figure 7. Classification maps of different classifiers on data set A. (a) Decision Tree. (b) Random Forest. (c) support vector machine (SVM). (d) XGBoost.
Figure 7. Classification maps of different classifiers on data set A. (a) Decision Tree. (b) Random Forest. (c) support vector machine (SVM). (d) XGBoost.
Sensors 18 00611 g007aSensors 18 00611 g007b
Figure 8. Classification maps of different classifiers on data set B. (a) Decision Tree. (b) Random Forest. (c) SVM. (d) XGBoost.
Figure 8. Classification maps of different classifiers on data set B. (a) Decision Tree. (b) Random Forest. (c) SVM. (d) XGBoost.
Sensors 18 00611 g008
Figure 9. Confusion matrices acquired via XGBoost on both data sets. (a) Data set A. (b) Data set B.
Figure 9. Confusion matrices acquired via XGBoost on both data sets. (a) Data set A. (b) Data set B.
Sensors 18 00611 g009
Figure 10. Effects of the training sample size. (a) Data set A. (b) Data set B.
Figure 10. Effects of the training sample size. (a) Data set A. (b) Data set B.
Sensors 18 00611 g010
Figure 11. Comparison of time costs on data sets A and B. (a) Data set A. (b) Data set B.
Figure 11. Comparison of time costs on data sets A and B. (a) Data set A. (b) Data set B.
Sensors 18 00611 g011
Figure 12. Classification results of different Q levels in SRM for the two data sets. (a) Data set A. (b) Data set B.
Figure 12. Classification results of different Q levels in SRM for the two data sets. (a) Data set A. (b) Data set B.
Sensors 18 00611 g012
Figure 13. Final classification map of data set A. (a) Superpixels generated by SRM. (b) Final classification map.
Figure 13. Final classification map of data set A. (a) Superpixels generated by SRM. (b) Final classification map.
Sensors 18 00611 g013
Figure 14. Final classification map of data set B. (a) Superpixels generated by SRM. (b) Final classification map.
Figure 14. Final classification map of data set B. (a) Superpixels generated by SRM. (b) Final classification map.
Sensors 18 00611 g014
Figure 15. Confusion matrices with selected Q levels for both data sets. (a) Dataset A. (b)Dataset B.
Figure 15. Confusion matrices with selected Q levels for both data sets. (a) Dataset A. (b)Dataset B.
Sensors 18 00611 g015
Figure 16. Images of roads surrounding the Jinyin Lake in data set A. a, b, and c are different images of roads. The images were taken on 16 August 2017.
Figure 16. Images of roads surrounding the Jinyin Lake in data set A. a, b, and c are different images of roads. The images were taken on 16 August 2017.
Sensors 18 00611 g016
Figure 17. Ground truth map and final classification map of data set A with four classes. (a) Ground trup map. (b) Final classification map.
Figure 17. Ground truth map and final classification map of data set A with four classes. (a) Ground trup map. (b) Final classification map.
Sensors 18 00611 g017
Figure 18. Confusion matrix for data set A with four classes.
Figure 18. Confusion matrix for data set A with four classes.
Sensors 18 00611 g018
Figure 19. Pauli RGB image (a) and ground truth map (b) of AIRSAR Flevoland data set.
Figure 19. Pauli RGB image (a) and ground truth map (b) of AIRSAR Flevoland data set.
Sensors 18 00611 g019
Figure 20. Classification results of different Q levels in SRM for AIRSAR Flevoland data set.
Figure 20. Classification results of different Q levels in SRM for AIRSAR Flevoland data set.
Sensors 18 00611 g020
Figure 21. Final classification map of AIRSAR Flevoland data set. (a) Superpixels generated by SRM. (b) Final classification map.
Figure 21. Final classification map of AIRSAR Flevoland data set. (a) Superpixels generated by SRM. (b) Final classification map.
Sensors 18 00611 g021
Figure 22. Confusion matrix with the selected Q level for AIRSAR Flevoland data set.
Figure 22. Confusion matrix with the selected Q level for AIRSAR Flevoland data set.
Sensors 18 00611 g022
Table 1. Data description of the used GF-3 PolSAR images.
Table 1. Data description of the used GF-3 PolSAR images.
ParameterImage #1Image #2
Imaging time29 May 20171 April 2017
PolarizationAHVAHV
Product typeSLCSLC
Imaging modeQPSIQPSI
Satellite directionAscendingAscending
Nominal resolution (m)88
Spatial resolution [Range × Azimuth] (m)5.53 × 2.255.20 × 2.25
Image size [Range × Azimuth] (pixel)7469 × 62107856 × 6805
Incidence angles ( ° )35.29–36.9933.69–35.61
Table 2. Classification accuracies on data set A: by classes and overall accuracy (%).
Table 2. Classification accuracies on data set A: by classes and overall accuracy (%).
ClassifiersWaterVegetationBuildingOA
Decision Tree98.9072.7676.6686.36
Random Forest99.2385.8475.3689.26
SVM99.6687.6972.3989.06
XGBoost99.1087.0276.9189.91
Table 3. Classification accuracies on data set B: by classes and overall accuracy (%).
Table 3. Classification accuracies on data set B: by classes and overall accuracy (%).
ClassifiersWaterBuildingFarmlandVegetation1Vegetation2OA
Decision Tree99.4270.3479.2781.6578.7082.49
Random Forest99.7676.5883.7587.3884.0386.64
SVM99.8372.2586.1289.9285.5487.96
XGBoost99.7872.9886.3788.5586.4388.15
Table 4. Q levels with the best overall accuracies (OAs) on data set A.
Table 4. Q levels with the best overall accuracies (OAs) on data set A.
Decision TreeRandom ForestSVMXGBoost
Best OA (%)88.1991.8191.3092.20
Q level203807973
Table 5. Q levels with best OAs on data set B.
Table 5. Q levels with best OAs on data set B.
Decision TreeRandom ForestSVMXGBoost
Best OA (%)92.6893.7794.6394.58
Q level86392395
Table 6. Detailed content of feature sets.
Table 6. Detailed content of feature sets.
Feature SetPolarimetric Features
FS1 | H H | 2 , | V V | 2 , T 11 , T 22 , T 33
FS2FS1, H, α , A
FS3FS2, | ρ |
FS4FS3, V P S , V P V , V P D
FS5FS3, Y P S , Y P V , Y P D , Y P C
FS6All the m = 16 features
Table 7. Accuracies by classes and OAs with different feature sets on data set A (%).
Table 7. Accuracies by classes and OAs with different feature sets on data set A (%).
Feature SetWaterVegetationBuildingOA
FS198.5476.9463.4483.42
FS298.8684.1570.8087.38
FS398.3181.4373.7887.32
FS499.3085.5878.4690.11
FS598.5683.5680.4389.84
FS699.1087.0276.9189.91
Table 8. Accuracies by classes and OAs with different feature sets on data set B (%).
Table 8. Accuracies by classes and OAs with different feature sets on data set B (%).
Feature SetWaterBuildingFarmlandVegetation1Vegetation2OA
FS199.3950.8579.1278.4380.9681.20
FS299.4369.7384.8086.8184.5586.60
FS399.3069.7385.7188.4082.1086.62
FS499.7774.5985.9587.6787.4088.19
FS599.4872.4186.7689.2985.2988.06
FS699.7872.9886.3788.5586.4288.15
Table 9. OAs of different classifiers on AIRSAR Flevoland data set.
Table 9. OAs of different classifiers on AIRSAR Flevoland data set.
Decision TreeRandom ForestSVMXGBoostLee [38]Tao [10]Qin [11]
Best OA (%)93.9091.8788.3193.3981.6394.5896.15
Q level253774477

Share and Cite

MDPI and ACS Style

Dong, H.; Xu, X.; Wang, L.; Pu, F. Gaofen-3 PolSAR Image Classification via XGBoost and Polarimetric Spatial Information. Sensors 2018, 18, 611. https://0-doi-org.brum.beds.ac.uk/10.3390/s18020611

AMA Style

Dong H, Xu X, Wang L, Pu F. Gaofen-3 PolSAR Image Classification via XGBoost and Polarimetric Spatial Information. Sensors. 2018; 18(2):611. https://0-doi-org.brum.beds.ac.uk/10.3390/s18020611

Chicago/Turabian Style

Dong, Hao, Xin Xu, Lei Wang, and Fangling Pu. 2018. "Gaofen-3 PolSAR Image Classification via XGBoost and Polarimetric Spatial Information" Sensors 18, no. 2: 611. https://0-doi-org.brum.beds.ac.uk/10.3390/s18020611

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop