Next Article in Journal
Nonlocal Tensor Sparse Representation and Low-Rank Regularization for Hyperspectral Image Compressive Sensing Reconstruction
Previous Article in Journal
Hyperspectral Anomaly Detection via Dictionary Construction-Based Low-Rank Representation and Adaptive Weighting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Image Classification Using Similarity Measurements-Based Deep Recurrent Neural Networks

1
Department of Geography, College of Geosciences, Texas A&M University, College Station, TX 77843, USA
2
Center for Geospatial Sciences, Applications and Technology, Texas A&M University, College Station, TX 77843, USA
3
Department of Computer Science and Engineering, Texas A&M University, College Station, TX 77843, USA
*
Author to whom correspondence should be addressed.
Submission received: 27 November 2018 / Revised: 28 December 2018 / Accepted: 16 January 2019 / Published: 19 January 2019
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
Classification is a common objective when analyzing hyperspectral images, where each pixel is assigned to a predefined label. Deep learning-based algorithms have been introduced in the remote-sensing community successfully in the past decade and have achieved significant performance improvements compared with conventional models. However, research on the extraction of sequential features utilizing a single image, instead of multi-temporal images still needs to be further investigated. In this paper, a novel strategy for constructing sequential features from a single image in long short-term memory (LSTM) is proposed. Two pixel-wise-based similarity measurements, including pixel-matching (PM) and block-matching (BM), are employed for the selection of sequence candidates from the whole image. Then, the sequential structure of a given pixel can be constructed as the input of LSTM by utilizing the first several matching pixels with high similarities. The resulting PM-based LSTM and BM-based LSTM are appealing, as all pixels in the whole image are taken into consideration when calculating the similarity. In addition, BM-based LSTM also utilizes local spectral-spatial information that has already shown its effectiveness in hyperspectral image classification. Two common distance measures, Euclidean distance and spectral angle mapping, are also investigated in this paper. Experiments with two benchmark hyperspectral images demonstrate that the proposed methods achieve marked improvements in classification performance relative to the other state-of-the-art methods considered. For instance, the highest overall accuracy achieved on the Pavia University image is 96.20% (using both BM-based LSTM and spectral angle mapping), which is an improvement compared with 84.45% overall accuracy generated by 1D convolutional neural networks.

Graphical Abstract

1. Introduction

Hyperspectral remote-sensing images (HSI) can entail both abundant spectral and spatial information, which generally provides enhanced capability of distinguishing different objects from one another, relative to multispectral images, and play an important role in a variety of research domains, such as precision agriculture [1], land-use monitoring [2,3], change detection [4,5], and environment measurements [6]. For such subfields, classification is a critical technology, where each pixel in an HSI will be classified with a pre-defined label, and encouraging achievements have been produced from different methods.
Based on the acquisition of training samples, HSI classification frameworks can be divided into three types: unsupervised, semi-supervised, and supervised classification. Considering the availability of training samples and classification performances, supervised methods are investigated the most and are typically selected as benchmark algorithms, such as k-nearest-neighbor [7,8,9], support vector machines (SVM) [10,11,12], sparse representation [13,14,15,16], and artificial neural networks (ANNs) [17,18]. Since acquiring sufficient labeled samples in the field can be quite difficult and time-consuming, semi-supervised methods [19,20,21,22,23] and tensor-based methods [24,25,26], which can obtain satisfying results only based on a limited number of training samples, also warrant many investigations. To overcome the shortcoming of limited training samples, researchers have also found that spatial information can be employed as crucial complementary and supportive features in spectral-spatial feature-combination frameworks to improve classification performance [27,28,29]. Various spatial feature-extraction methods have been proposed for HSI classification; however, such kinds of spectral-spatial features still refer to low-level features [30], and the classification performances arising from spectral-spatial methods can be easily affected by the curse of dimensionality problem and the difficulty in discriminating among some classes. Therefore, methods for extracting more robust and representative features from the HSI image itself still need to be investigated.
In the most recent decade, deep learning (DL) has demonstrated marked effectiveness and robustness on large datasets. Considering the inherent deep architecture, which can be viewed as an efficient feature-extraction model, from low-level feature to the highly-abstract level, DL algorithms are always employed as end-to-end classifiers for both pixel- and image-based classification tasks in remote-sensing communities [31]. Chen et al. proposed a stacked autoencoder (SAE)-based HSI classification framework [32], and it was the first paper on DL applications in HSI processing. During unsupervised feature extraction, spatial information was incorporated in order to enhance the classification performance. The authors have also investigated the performance of deep belief network (DBN), another unsupervised deep feature learning method for HSI classification [33]. Additionally, even considering the relatively limited availability of ground-reference or other training data, supervised DL algorithms have also been well-explored and have posted generally impressive results. Convolutional neural networks (CNNs) can been viewed as a milestone in the DL epoch since its first successful application in target detection [34]. CNNs are similar to conventional ANNs in that they are all composed of neurons, activation functions, and loss functions for optimization. The most significant development pertaining to CNNs is that with CNNs, 2-dimensional (2D) images can be fed directly into the network architecture via the input layer by introducing convolutional layers, instead of transforming 2D images to one-dimensional (1D) vectors. Such a property makes image-based processing more efficient and straightforward, and applying a convolutional layer can utilize spatial contextual information with a specific receptive domain. However, regarding pixel-based processing of HSIs, CNNs cannot be employed directly due to its 2D filter-processing characteristic, and preprocessing is needed, including patch extraction. Hu. et al. [35] applied a 1D CNN (1DCNN) for pixel-based HSI classification, where a single pixel is considered as a 2D image whose height is equal to 1. Makantasis et al. [36] proposed a 2D CNN (2DCNN)-based HSI classification method where randomized principal component analysis (R-PCA) was conducted as a dimensionality reduction to reduce the number of 2D convolutional filters. Li et al. [37] investigated the effectiveness of a three-dimensional (3D) CNN (3DCNN)-based HSI classification technique, where spectral and spatial features were simultaneously exploited without any dimensionality reduction pre-processing.
Recurrent neural networks (RNNs) [38], another novel DL architecture with outstanding adaptability for handling sequence data has recently achieved promising performance, particularly for natural language processing (NLP) [39,40]. Since multi-temporal information can be obtained conveniently with the development of modern satellite remote-sensing technology, RNNs and an improved type of RNN, referred to as long short-term memory (LSTM) [41], were explored to extract temporal features from multiple images. Ienco et al. [42] utilized RNN and LSTM to perform land-cover classification on multi-temporal satellite images. In [43], Sharma et al. proposed a patch-based RNN framework incorporating both spectral and spatial information within a local window to classify Landsat 8 images. Furthermore, single-image-based RNN methods are also applied to HSIs. Mou et al. [44] proposed a novel RNN-based HSI classification algorithm by using a parametric rectified hyperbolic tangent function ( P R e t a n h ). In this framework, each individual pixel in the HSI can be regarded as one sequential feature for the RNN input layer. Wu et al. [45] investigated the combination of CNN and RNN layers on the spectral feature domain and employed the convolutional RNN (CRNN) model for HSI classification. The utilization of a CNN can extract patch-level local invariant information among spectral bands, which provides spatial contextual features for the following RNN layers. Shi et al. [46] proposed another strategy to design the sequential data in RNN model instead of taking spectral vector from all bands as one sequential data, but taking advantage of spatial neighbors. For this method, local spectral-spatial features were first extracted by exploiting a 3DCNN on a local image patch, and then sequences were built based on an eight-directional construction.
Although the aforementioned RNN-based DL models have significantly contributed to HSI processing efforts, there are still some critical problems that need to be addressed. The first issue is the limitation of training sample. Acquiring sufficient labeled training data for HSI classification is often difficult and time consuming. Moreover, satisfactory DL-based classification accuracy has always relied upon very large sets of training samples. Therefore, obtaining convincing HSI classification results by utilizing limited training data for DL models is a challenging task. However, unlabeled samples, which are relatively easier to acquire than labeled samples, have already been investigated for HSI classification purposes under semi-supervised classification frameworks [19,20,21,22]. Such investigations illustrate the potential effectiveness of unlabeled data for such a purpose. Another critical issue involves the construction of sequential data for the RNN model. In [44,45], the respective authors analyzed the HSI from the perspective of a sequential point, meaning that each pixel is considered to be a data sequence since all pixels in the HSI are sampled densely from the entire spectrum, and they are expected to have dependencies between different bands. Nonetheless, such dependencies still need to be explored in order to more fully exploit the integrity of the full spectral signature. In order to distinguish different classes, it is frequently advantageous to utilize the information encapsulated within the entire reflectance spectrum, as is the case with many conventional classification methods. Furthermore, exploiting the spectral feature directly in the RNN model will introduce more parameters that need to be computed and optimized in the training step.
In this paper, we propose a novel LSTM-based HSI classification framework with spatial similarity measurements (SSM), which is inspired by [47], where the LSTM model and the spatial location are combined simultaneously. First, the sequential feature for each pixel is constructed by selecting candidates from the whole image based on the similarity between each candidate and the target pixel. This selection method mainly relies upon two different similarity measurements where spectral and spatial information are considered, namely pixel-matching-based (PM) spatial similarity measurements and block-matching-based (BM) spatial similarity measurements, respectively. LSTM entails the significant capability of handling sequential data, and it achieves outstanding performance in NLP. The proposed similarity-measuring strategies provide an innovative framework to extract sequential features for HSI classification by employing all pixels in the entire HSI, regardless of whether the pixel candidates to be a given sequential feature are labeled or not. The proposed sequential feature effectively encodes the dependency of the target pixel with regard to its contexts, where the pixel-level spectral similarities and the patch-, or block-level contextual similarities are naturally encoded, respectively. More specifically, LSTM assumes that closer “time steps” (which denote selected pixels/blocks with higher similarity) in general have stronger feature sharing, while also allowing for longer-term dependency that can account for non-local similarity and long-tail effects. The motivation for extending one pixel to its sequential features is essentially to find a new feature embedding of the pixel, which makes reference to other similar pixels. The action of re-ordering those pixels in terms of their similarity measures is to ensure that all obtained sequential features admit “comparable” formats in terms of monotonically-decreasing similarity to the original pixel (i.e., the target pixel, or the given pixel of interest). In this framework, the influence of unlabeled data in an HSI is enhanced compared with conventional supervised-learning methods due to our proposed spatial selection, where any pixel can be selected as a candidate to construct sequential features. In view of global searching based on the whole image, more supportive pixels are incorporated with a wider receptive domain. In addition, spatial contextual information, which has already been utilized to reduce the “salt-and-pepper” phenomenon in remote-sensing/HSI classification [48], is also investigated here in the BM-based framework. Similarity measurements of two pixels is implemented by using the neighboring points of two pixels instead of their spectral feature vectors. Compared with the PM-based method, the BM-based scheme can obtain more typical sequential feature representation combining both spectral and spatial features together. In summary, this is the first study to propose such methods operating over the entire image. This is important because these new, novel methods can incorporate additional information collected throughout the whole image by exploiting unlabeled pixels, instead of utilizing only the limited prior information in the form of labeled pixels. Figure 1 illustrates the framework of our proposed method.
The organization of the remainder of this paper is as follows: Section 2 provides a brief introduction to the original RNN and LSTM models. Section 3 describes the proposed LSTM classification framework based on pixel matching and block matching. Experimental results are discussed in Section 4. Section 5 presents the conclusions.

2. Background: RNN and LSTM

2.1. RNN

The recurrent neural network (RNN) has shown great capability in time-sequence data processing, including NLP [49] and speech recognition [40]. A significant characteristic of a time sequence is that there is typically a strong relationship between a given sample and the previous samples. In the hidden Markov model (HMM), which is a widely-utilized sequence model in language processing, the probability of a specific state depends only on its previous state, instead of all previous states. Let X = [ x 1 , x 2 , , x t ] be the sequence data where t is the label of state. x 1 represents the data at the first state, and x t represents the data at the t state. The Markov assumption can be formulated as
P ( x t | x 1 , , x t 1 ) = P ( x t | x t 1 ) ,
where P ( · ) expresses the conditional probability. Compared with HMM, RNN is quite similar with HMM since the computation on the current state relies on the previous state. In contrast to conventional ANNs, the RNN has a circular processing on the sequential data, which means that such processing will be applied on each data instance in the sequence, and the result at each state relies on the previous state. This circular processing also represents the parameter sharing. Parameter sharing is a prevalent method to control the number of parameters in a DL scheme. Still given X = [ x 1 , x 2 , , x t ] as sequential data, the hidden state s t can be represented as
s t = f s ( W xs x t + W ss s t 1 + b t ) ,
where W xs is the weight matrix from input data to the hidden state, and W ss is the weight matrix from the current state to the next state, respectively. b t is the bias variable. s t denotes the hidden state at time step t, and f s ( · ) represents the nonlinear activation function. The calculation of the output at state t is quite similar with Equation (2) as follows:
y t = f y ( W sy s t + b y ) ,
where W sy is the weight matrix from the hidden state to the output. b y is the bias, and f y ( · ) is the nonlinear activation function.
The hidden state s t can be viewed as the memory of the RNN model, as it is calculated based on the previous state through forward propagation. Meanwhile, sequential data in the previous states are taken into consideration as well. In such forward propagation, some parameters, including three different weight matrices W xs , W ss , and W sy , are shared across all steps, which is quite different from a traditional neural network. The parameter-sharing scheme reduces the number of trainable parameters, and makes the total computation more efficient.

2.2. LSTM

In Equation (2), the calculation of the hidden state depends on the previous state. However, with the increase in the length of the sequence data, gradient vanishing and gradient exploding will be introduced in this recurrent model due to the forward and backward propagations of weight matrices. To address this issue, long short-term memory was developed with a more sophisticated recurrent neuron. In LSTM, each recurrent neuron can be regarded as a cell state. Similar to the conventional RNN, LSTM also employs the previous state as the input to the current state. However, with LSTM, there are three gates, including forget gate, update gate, and output gate, to control the update of the current neuron. Figure 2 illustrates the basic structure of the LSTM recurrent unit.
The first part of LSTM is the forget gate which determines if the previous state will be retained or not. Still given X = [ x 1 , x 2 , , x t ] as sequential data, y t and s t are the output and the hidden state at step t, respectively. The common computation for forget gate f t is as follows:
f t = σ f ( W yf y t 1 + W xf x t + b f ) ,
where the W ( · ) terms denote the weight matrices, and the b ( · ) term is the bias variable. σ ( · ) is the logistic sigmoid function. The following step is to compute the update gate u t and a new candidate state value s t ˜ :
u t = σ u ( W yu y t 1 + W xu x t + b u ) ,
s t ˜ = t a n h ( W ys y t 1 + W xs x t + b s ) ,
where t a n h ( · ) is the hyperbolic tangent function. Then, the new hidden step s t can be updated by using the aforementioned equations
s t = s t 1 × f t + s t ˜ × u t .
Finally, the output gate and the output of the current neuron yields:
o t = σ o ( W yo y t 1 + W xo x t + b o ) ,
y t = o t × t a n h ( s t ) .

3. Spatial Similarity Measurements in LSTM

For the LSTM model, the sequential feature is a critical issue when training LSTM since determining the representative feature will improve classification performance and reduce the training-time cost. In this section, the spatial similarity measurement-based LSTM model will be introduced as a method to construct sequential features. First, two different strategies utilized in SSM will be discussed, named PM-based and BM-based schemes. For each of them, when computing the similarity between pixels, two distance measurements are investigated, which are Euclidean distance (EU) and spectral angle mapper (SAM). Furthermore, we will introduce the way of constructing sequence as the input of LSTM model.

3.1. Pixel Matching

Measuring the similarity in the pixel data vectors between different pixels is a common technology in many HSI analysis applications, such as endmember-based analysis [50], manifold learning [51,52], and graph-based semi-supervised learning [22]. Since HSIs have abundant spectral information, typically entailing hundreds of bands, spectral features collected from all bands have the most discriminative capability to distinguish different ground objects or materials encompassed within the given image, and have been most widely utilized in HSI classification [53]. In the pixel-matching scheme, pairwise spectral similarity measurements are applied for all pixels. Suppose we have HSI data X R L × C × B with L rows, C columns, and B bands. X can be rewritten as X = [ x 1 , x 2 , . . . , x N ] R B × N with row-major order, where N is the total number of pixels in the HSI, which equals L C . For any pixels x i in X , the distances measured between x i and all pixels of X will be computed as follows:
d PM ( x i , X ) = [ d ( x i , x 1 ) , , d ( x i , x j ) , , d ( x i , x N ) ] ,
where d ( x i , x j ) denotes the distance between x i and x j , and d ( · ) is the distance calculation function. There are multiple methods available to compute pairwise distance. In this paper, Euclidean distance and the spectral angle mapper (SAM) are utilized in the pixel-matching scheme. Euclidean distance is a well-known distance measurement, and its definition is given as follows:
d EU ( x i , x j ) = ( x i 1 x j 1 ) 2 + ( x i 2 x j 2 ) 2 + + ( x i B x j B ) 2 = x i x j .
The other distance measure adopted in this study is SAM, which is investigated well in endmember-based HSI classification. It can be defined as follows:
d SAM ( x i , x j ) = cos 1 ( x i T x j x i x j ) .
In the following text of this paper, we use d ( · ) ( ) to represent either the EU or SAM distance calculation function in the pixel-wise measurement. Therefore, Equation (10) can be rewritten as follows:
d PM ( x i , X ) = [ d ( · ) ( x i , x 1 ) , , d ( · ) ( x i , x j ) , , d ( · ) ( x i , x N ) ] .

3.2. Block Matching

Although spectral features provide rich, significant information that facilitates discrimination of different ground objects in HSI, classification accuracy based on utilization of spectral features alone is not always satisfactory due to the “salt-and-pepper” phenomenon [48]. Given Tobler’s First Law of Geography [54], incorporation of spatial contextual information has attracted increasing attention in the research literature in recent years and has exhibited the capability to reduce “salt-and-pepper” noise and improve classification performance, including yielding smoother classification maps. In the current study, we utilize the image patch distance (IPD), proposed in [55], for similarity measurement that considers local spatial information. The IPD method was developed based on the Hausdorff distance [56], also referred to as the Pompeiu–Hausdorff distance. Instead of using spectral features to measure the pairwise similarity between two pixels, spatial neighbors within the respective local window of such two pixels will also be employed. Let w be the local window size, and the s i represents the block neighborhood, where pixel x i is centered within the w × w spatial window. All block sets S can be defined as follows:
S = [ s 1 , s 2 , , s N ] .
Given s i , s j S , we first calculate the distances between one arbitrary pixel x m from s i and s j and then select the minimum distance as follows:
d m i n ( x m , s j ) = min x n s j d ( · ) ( x m , x n ) ,
where d ( · ) is the distance-measuring function. Correspondingly, the minimum distance between one arbitrary pixel x m from s j and s i can be computed in the following manner:
d m i n ( x m , s i ) = min x n s i d ( · ) ( x m , x n ) .
Therefore, the definition of block matching between s i and s j is:
d BM ( s i , s j ) = m = 1 w 2 ( max ( d m i n ( x m , s i ) , d m i n ( x m , s i ) ) ) = m = 1 w 2 ( max ( min x n s j d ( · ) ( x m , x n ) , min x n s i d ( · ) ( x m , x n ) ) ) .
Finally, the ultimate BM measurements are defined as:
d BM ( x i , X ) = d BM ( s i , X ) = [ d BM ( s i , s 1 ) , , d BM ( s i , s j ) , , d BM ( s i , s N ) ] .

3.3. Sequential Feature Extraction

After measuring the pairwise similarity among pixels in the whole image, the sequential feature for each pixel is extracted so that it can be fed into the RNN model directly. Based on the aforementioned two matching schemes, given one pixel x i X , its corresponding matching vector is
d M ( x i , X ) = [ d M ( x i , x 1 ) , , d M ( x i , x j ) , , d M ( x i , x N ) ] ,
where d M ( · ) denotes the pairwise matching function introduced in previous sections. Note that the order of d M ( x i , X ) is determined only based on its location in the image. To characterize a representative sequential feature, d M ( · ) is reordered based on the degree of similarity, and then the corresponding pixels are selected as the sequential representation. With the definition given in Equation (19), we assume we have one pixel x i X , its reordered d M ( x i , X ) , named d s f ( x i , X ) , is built as follows:
d s f ( x i , X ) = [ d s f i 1 , , d s f i j , , d s f i N ] ,
where d s f i j denotes the distance-measuring result between x i and x j . d s f ( x i , X ) is the ascending sort of d M ( x i , X ) . More specifically, d s f 1 is the minimum value among d s f ( x i , X ) , and d s f 2 is the second-smallest value. During ascending sorting, pixels most similar to x i among all pixels in the whole image will be selected as the sequential representation of x i . Note that not all candidates in d s f ( x i , X ) will be considered, and the parameter l is defined to control how many candidates can be selected or to determine the length of such sequence. The first l pixels with distance-measuring results from d s f 1 to d s f l will be selected, and the first elements of this sequence is x i itself since d M ( x i , x i ) equals zero, which is the minimum value in d M ( x i , X ) . Given the sequence length l, the final sequential feature of x i can be defined:
d s f ( x i ) = [ x sf i 1 , , x sf ij , , x sf il ] ,
where x sf ij represents x j , whose distance-measuring result is located at the jth place.

4. Experimental Setup, Results, and Discussion

4.1. Datasets

In this study, two benchmark HSI datasets were utilized, including Pavia University, and Salinas images, as displayed in Figure 3 and Table 1. The Pavia University image was collected by the Reflective Optics System Imaging Spectrometer (ROSIS) sensor. It consists of 102 spectral bands, with a spectral range from 430 nm to 860 nm. The image spatial resolution is 1.3 m, and the total image size is 610 × 340 pixels. For the Pavia University image extent, nine (9) classes were considered in the classification experiments. The Salinas image was acquired via the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS), and the image contains 512 lines × 217 samples, with the spatial resolution of 3.7 m. After removing 20 noise and water-absorption bands, 204 spectral bands remained for subsequent analysis. The ground-reference data for the Salinas image entails 16 classes.

4.2. Experimental Design

To evaluate the performance of the proposed SSM-based LSTM methods, three algorithms, including SVM, 1DCNN [35], and 1DLSTM [44], are investigated as baseline algorithms. For SVM, the radial basis function (RBF) is utilized as kernel function, and the parameters of SVM are acquired by cross validation. For the following two deep-learning algorithms, they are 1D-based architectures, where spectral features are fed into the classifier directly. For the 1DCNN, two convolutional layers, two max pooling layers, and one fully-connected (FC) layer are selected due to the limited available training samples. For the 1DLSTM architecture, three LSTM layers and one FC layer are adopted. Different model structures are implemented based on different images, and the specific parameter settings for the Pavia Univeristy and Salinas images are summarized in Table 2, where the convolutional layer is represented as “Conv(number of kernels)-(kernel size)”, maxpooling layer is performed as “Maxpooling-(kernel size)”, and LSTM layer is denoted as “LSTM-(kernel size)”. Regarding the proposed methods, both pixel-matching and block-matching are investigated, and, for each matching scheme, EU and SAM are employed as distance measurements. Therefore, there are four different LSTM-based classification frameworks investigated here, named LSTM_PM_EU, LSTM_PM_SAM, LSTM_BM_EU, and LSTM_BM_SAM. For the LSTM structure, we use four recurrent layers and two fully-connected layers, and the length of the sequential feature is 20. The size of the recurrent layers are 32, 64, 128, and 256, respectively, and the size of first fully-connected layer is 50. The second fully-connected layer is applied for the purpose of classification, and its length equals the number of classes. During training of the recurrent model, the batch size is set to 20, and the number of epochs is 500.
In order to evaluate classification performance quantitatively, all ground-reference data for each image is randomly split into training and testing sample sets. In our experiments, we randomly select 200 samples per class as training data, with the remaining ground-reference data used as testing data. Ten replications of the experiments with such random selections were performed, and all classification accuracies were averaged across the ten replications. Furthermore, three other quantitative indicators were also adopted for the evaluation, including overall accuracy (OA), average accuracy (AA), and the Kappa coefficient (Kappa) [57]. The pixel-matching and block-matching experiments were implemented on the Texas A&M High Performance Research Computing (HPRC) system, and the remaining experiments, such as training LSTM models and classification accuracy assessments, were carried out on a local workstation with a 3.2 GHz Intel(R) core i7-8700 Central Processing Unit (CPU), and a NVIDIA(R) GeForce GTX 1070 graphics card.

4.3. Classification Results: Pavia University Image

The first set of experiments is conducted on the Pavia University image. The quantitative results are shown in Table 3, where values in bold are the highest class-specific accuracies and the standard deviations are also presented, which are calculated based on ten OAs obtained from the aforementioned ten (10) experimental replications. The classified images are displayed in Figure 4 for qualitative analysis, which are obtained from the fifth trial. As shown in Table 3, the block-matching-based method LSTM_BM_SAM achieved best performance, with 96.20% OA, 94.65% AA, and 94.91% Kappa. For the first three benchmark algorithms, the highest OA (i.e., 84.45%) is obtained from 1DCNN. Regarding our newly-proposed pixel-matching-based LSTM frameworks, the OA of LSTM_PM_SAM is 84.56%, exhibiting limited improvement relative to SVM, 1DCNN, and 1DLSTM, and the classification performance of LSTM_PM_EU even decreases relative to 1DCNN and 1DLSTM. However, after incorporating spatial information via similarity measurements, LSTM_BM_EU and LSTM_BM_SAM obtain marked improvements over all non-block-matching methods, with 95.96% and 96.20% OA, respectively. Within each matching method, the performance of SAM is always better than that of the Euclidean distance. Regarding the AA of each class, class 7 (Bitumen, Red) is more difficult to discriminate compared with other classes due to the mixed spectral features. The proposed LSTM_BM_SAM improves the original SVM OA by more than 35%.
From the classification maps shown in Figure 4, marked improvements in classification performance are visually apparent. In Figure 4b–f, “salt-and-pepper” noise is still obvious due to the lack of incorporation of spatial contextual information in the classification. Within the red-rectangle annotation, many class 2 (Meadow, Bright Green) pixels are misclassified as class 6 (Bare Soil, Yellow), and class 3 (Gravel, Brown), as shown in Figure 4b–f. However, the classification maps derived from LSTM_BM_EU and LSTM_BM_SAM (Figure 4g,h) are spatially smooth and generally correctly classified, where most discrete, spurious/misclassified points are eliminated if they are located within an otherwise homogeneous area. Therefore, combining spatial contextual information can yield marked alleviation of image misclassification. Similar to what is observed within the red-rectangle annotation, more accurate and homogeneous classification results can be achieved within the red-circle annotation as well. Such results demonstrate the validity and capability of combining spatial and spectral features together when measuring the similarity between two pixels, and the effectiveness of constructing a sequential feature for a specific pixel based on such similarity between that target pixel itself and candidates from the whole image.

4.4. Classification Results: Salinas Image

For the Salinas image, the results are similar to those attained and described in Section 4.3. The quantitative results are shown in Table 4, where, again, values in bold are the highest class-specific accuracies. The OAs of the SAM-based method are lower than those associated with its corresponding Euclidean distance-based method, and the performance of the block-matching strategy is always better (more accurate) than that of the pixel-matching scheme, where spatial contextual information is ignored. The best classification performance is still obtained from LSTM_BM_SAM, with OA = 90.63%, AA = 93.95%, and Kappa accuracy = 89.55%. Class 15 (Vineyard_untrained, Violet) is the class with the lowest accuracy due to the high spectral and thematic similarity between this vineyard class and other grape fields, and the best classification result for this class is acquired from 1DCNN (among SVM, 1DCNN, and 1DLSTM), with 59.34% accuracy. However, LSTM_BM_EU and LSTM_BM_SAM markedly improve classification accuracy by utilizing spatial features, where class 15 accuracies increase by 7.99% and 10.51%, respectively, compared with the 1DCNN result.
Manual interpretation of the classification maps shown in Figure 5 enables us to determine why class 15 (Vineyard_untrained) entails the lowest OA. Note that many class 15 (Vineyard_untrained, Violet) pixels are misclassified as class 8 (Grapes_untrained, Baby Blue). Pixel-matching-based methods, including LSTM_PM_EU and LSTM_PM_SAM, still yield much discrete noise within the red-circle annotation in Figure 5. However, LSTM_BM_EU and LSTM_BM_SAM produce more homogeneous and smoothed classification results for the Grapes_untrained class, especially within the red-circle annotation, illustrated in Figure 5g,h. Within the red-rectangle annotation, we can see that it is difficult to classify class 10 (Corn_senesced_green_weeds, Brown), for example, and many pixels are misclassified in Figure 5b–f. However, such misclassification is markedly minimized when applying block-matching-based methods, i.e., LSTM_BM_EU (Figure 5g) and LSTM_BM_SAM (Figure 5h). Such improvement illuminates the advantages of utilizing spatial contextual information when measuring the pixel-wise distances, especially when it is challenging to discriminate between two classes with very similar spectral features.

4.5. Parameter Sensitivity Analysis

The influence of different parameter values associated with our proposed methods is investigated in this section, including the length of sequential feature l, and the size of the local window w, utilizing block-matching-based methods. The effect of varying the value of l is tested on LSTM_PM_EU, LSTM_PM_SAM, LSTM_BM_EU, and LSTM_BM_SAM, and the effect of varying the value of w is tested using LSTM_BM_EU and LSTM_BM_SAM.
For the first parameter, l, five different lengths (10, 20, 30, 40, and 50) are investigated, while the window size utilized in LSTM_BM_EU and LSTM_BM_SAM is fixed at 5. The results are shown in Figure 6. Regarding the Pavia University data, the best performances for those four proposed methods are obtained with different sequential feature lengths (Figure 6a). Sequential feature lengths of 20 and 10 result in the highest classification OAs for LSTM_PM_EU (82.70%) and LSTM_PM_SAM (86.68%), respectively. For the two block-matching-based methods, the highest OA for LSTM_BM_EU is 96.54% (when l is 50), and setting l to 20 yields the highest-accuracy result for LSTM_BM_SAM. For the pixel-matching-based algorithms, the classification performance of the Euclidean-distance measure is always better than that of SAM. Furthermore, selecting a smaller sequential length (i.e., 10 or 20) is suitable for these two methods. Regarding the block-matching-based methods, Euclidean distance performs better than SAM, except for l is 20. Smaller sequential lengths used with SAM result in higher OAs, but such lengths are not suitable when using the Euclidean distance measure. Nevertheless, the difference in the resultant classification accuracies when employing these two distance measurements in the block-matching scheme is smaller than what it is in pixel-matching scheme. Moreover, the standard deviations for those four methods, across the five sequential lengths, are 1.0020, 1.4280, 0.5342, and 0.4805, for LSTM_PM_EU, LSTM_PM_SAM, LSTM_BM_EU, and LSTM_BM_SAM, respectively, which illustrates that, for the Pavia University data, the block-matching scheme is less sensitive than the pixel-matching method to the sequential-length parameter value.
For the Salinas data, the best choices for l vary depending on the algorithm. As shown in Figure 6b, the length of 40 yields highest accuracies in LSTM_PM_EU and LSTM_BM_SAM. LSTM_PM_SAM achieves best OA when utilizing 20 as the sequential length. Length of 30 results in the highest OA in LSTM_BM_EU. Different from what we observed from Figure 6a, within the pixel-matching-based schemes, SAM always performs better than Euclidean distance except l = 20. Additionally, smaller length provides a better performance for LSTM_PM_SAM (l = 20) but not applicable in LSTM_PM_EU. For the blocking-matching schemes, SAM is the more robust distance measurement since it performs better at four lengths (10, 20, 40, 50), and obtains the highest accuracy with l = 40. Euclidean distance only yields better result compared with SAM at the length of 30 and that is the highest OA among all lengths. For the standard deviations of those four methods, they are 0.3658, 0.4739, 0.9334, and 0.8140, respectively, which exhibits that pixel-matching-based methods is less sensitive than block-matching-based ones. However, due to the higher OAs obtained by LSTM_BM_EU and LSTM_BM_SAM, block-matching schemes are still the applicable methods to classify the Salinas data.
Another investigation of parameter l is its influence on the training time of the LSTM model since a larger l will introduce more parameters to be learned in the LSTM model and will result in more processing time. The training times of different approaches are given in Table 5. Those training times are the average values obtained from the 10 replications. Different methods with the same l have similar training times for both the Pavia University and Salinas images. However, the training time differs when a different l is applied within one LSTM model. As an example, consider the application of the LSTM_BM_SAM to the Pavia University image; the training time is 32.00 min when l is 10. The training time increases along with utilization of larger l, and it reaches 142.07 min, which is more than four times the minimum training time consumption. Fortunately, BM-based methods are less sensitive compared with PM-based methods regarding the selection of different l, which can be obtained from Figure 6. To balance the computation time cost and classification performance, choosing a smaller l (e.g., 10 or 20) is an appropriate strategy for our proposed methods, even though PM-based methods are relatively more sensitive with respect to parameter l.
For the second parameter, w, four different window sizes ( 5 × 5 , 7 × 7 , 9 × 9 , 11 × 11 ) and two methods (LSTM_BM_EU and LSTM_BM_SAM) are chosen for the comparison experiments, where l is fixed at 20. These results are shown in Figure 7. For the Pavia University data-based results (Figure 7a), the overall classification accuracy of LSTM_BM_SAM is higher than that of LSTM_BM_EU. LSTM_BM_EU obtains a higher OA only when w is set at 7, and that is also the best accuracy across all four window sizes. LSTM_BM_SAM achieves the best performance with a window size of 5 (the smallest window size considered), and the OA decreases as the window size increases. Regarding the Salinas data-based results, given in Figure 7b, the classification accuracies for LSTM_BM_SAM are still higher than those for LSTM_BM_EU, and both of those two methods realize the most accurate results with a window size of 5. Compared with parameter l, the optimal value for w is easier to determine in order to achieve a more accurate classification result. It is predictable that incorporating local spatial contextual information is likely to help better measure the similarity between two pixels, yielding improved classification performance. However, with increasing window size, too many neighboring pixels are included in the calculation, resulting in over-smoothed classification maps, and class spatial boundaries are not preserved. As a consequence, selection of a relatively small window size should introduce sufficient—though not too much—spatial information, leading to higher classification accuracies.

5. Conclusions

In this paper, we propose a novel LSTM HSI classification framework, where unlabeled data are well-exploited in order to construct sequential features from a single HSI. Instead of using spectral features as the sequential data structure of LSTM, similar pixels collected from the entire image are used to construct the respective sequential features. Specifically, when constructing a sequential feature, the similarity between a target pixel and all other pixels in the image is considered. To better depict the similarity between two pixels, two similarity-measuring strategies—pixel-matching and block-matching—are adopted here, where individual spectral features are utilized in the pixel-matching-based schemes, and both spatial and spectral information are employed in block-matching-based schemes. Such schemes take full advantage of unlabeled data in the HSI, as labeled data are almost always limited in nature and difficult to acquire for HSI classification. Moreover, block-matching-based schemes also consider spatial contextual information in the classification process, and it is demonstrated in this research that such schemes are effective in increasing HSI classification. Our proposed methods produce markedly more accurate results when operating on two well-known, extensively-studied HSI datasets compared with other selected baseline algorithms. Particularly regarding the Pavia University image, the LSTM_BM_SAM achieves the best classification performance, with 96.20% OA, which is 11.75% higher than the best result obtained by the three benchmark algorithms, which in this case was 1DCNN, with 84.45% OA. Furthermore, that OA is also higher than those from other three proposed methods (LSTM_PM_EU, LSTM_PM_SAM, and LSTM_BM_EU), with an OA increase of 13.50%, 11.64%, and 0.24%, relative to those respective methods. Additionally, in these experiments, BM-based methods always yield better results compared with their corresponding PM-based methods, which demonstrates the effectiveness and capability of the utilization of spatial contextual information.
Regarding the proposed block-matching method, fixed window sizes are applied for classification. In the future, we will explore adaptive window-size applications, intended to eliminate the phenomenon of over-smoothing in the classified images and to preserve the respective boundaries between different classes. In addition, measuring pixel-wise similarity from the entire HSI more efficiently still needs to be investigated in future research.
The proposed methods in this study combine similarity measurements and recurrent neural networks, and, although in the present study we focus on encoding spatial contextual information, future work may involve implementing these methods in a temporal context (i.e., in a true multi-temporal remote-sensing context).

Author Contributions

Z.W. and A.M. designed the study. A.M. and A.M.F. planned the experiments. Z.Y. designed the parallel computing experiments. A.M. ran all the experiments. A.M. wrote the draft. A.M.F. and Z.W. revised the draft.

Funding

This research was supported by NASA Land Cover/Land Use Change (NNH15ZDA001N-LCLUC: Land Cover/Land Use Change) Grant No. NNX17AK14G. The open access publishing fees for this article have been covered by the Texas A&M University Open Access to Knowledge Fund (OAKFund), supported by the University Libraries and the Office of the Vice President for Research.

Acknowledgments

The authors thank the four anonymous reviewers for their constructive remarks, which improved the clarity and quality of this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Haboudane, D.; Miller, J.R.; Pattey, E.; Zarco-Tejada, P.J.; Strachan, I.B. Hyperspectral vegetation indices and novel algorithms for predicting green LAI of crop canopies: Modeling and validation in the context of precision agriculture. Remote Sens. Environ. 2004, 90, 337–352. [Google Scholar] [CrossRef]
  2. Camps-Valls, G.; Tuia, D.; Bruzzone, L.; Benediktsson, J.A. Advances in hyperspectral image classification: Earth monitoring with statistical learning methods. IEEE Signal Process. Mag. 2014, 31, 45–54. [Google Scholar] [CrossRef]
  3. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  4. Eismann, M.T.; Meola, J.; Hardie, R.C. Hyperspectral change detection in the presenceof diurnal and seasonal variations. IEEE Trans. Geosci. Remote Sens. 2008, 46, 237–249. [Google Scholar] [CrossRef]
  5. Wang, Q.; Yuan, Z.; Du, Q.; Li, X. GETNET: A General End-to-End 2-D CNN Framework for Hyperspectral Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2018, 57, 1–11. [Google Scholar] [CrossRef]
  6. Plaza, A.; Du, Q.; Chang, Y.L.; King, R.L. High performance computing for hyperspectral remote sensing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2011, 4, 528–544. [Google Scholar] [CrossRef]
  7. Ma, L.; Crawford, M.M.; Tian, J. Local manifold learning-based k-nearest-neighbor for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2010, 48, 4099–4109. [Google Scholar] [CrossRef]
  8. Jia, X.; Richards, J.A. Fast k-NN classification using the cluster-space approach. IEEE Geosci. Remote Sens. Lett. 2005, 2, 225–228. [Google Scholar] [CrossRef]
  9. Wang, Q.; Zhang, F.; Li, X. Optimal Clustering Framework for Hyperspectral Band Selection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1–13. [Google Scholar] [CrossRef]
  10. Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
  11. Gualtieri, J.; Chettri, S. Support vector machines for classification of hyperspectral data. In Proceedings of the IEEE 2000 International Geoscience and Remote Sensing Symposium (IGARSS 2000), Honolulu, HI, USA, 24–28 July 2000; Volume 2, pp. 813–815. [Google Scholar]
  12. Tarabalka, Y.; Fauvel, M.; Chanussot, J.; Benediktsson, J.A. SVM and MRF-based method for accurate classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2010, 7, 736–740. [Google Scholar] [CrossRef]
  13. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification using dictionary-based sparse representation. IEEE Trans. Geosci. Remote Sens. 2011, 49, 3973–3985. [Google Scholar] [CrossRef]
  14. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Hyperspectral image classification via kernel sparse representation. IEEE Trans. Geosci. Remote Sens. 2013, 51, 217–231. [Google Scholar] [CrossRef]
  15. Zhang, H.; Li, J.; Huang, Y.; Zhang, L. A nonlocal weighted joint sparse representation classification method for hyperspectral imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2056–2065. [Google Scholar] [CrossRef]
  16. Wang, Q.; He, X.; Li, X. Locality and Structure Regularized Low Rank Representation for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2018, 1–13. [Google Scholar] [CrossRef]
  17. Benediktsson, J.A.; Swain, P.H.; Ersoy, O.K. Conjugate-gradient neural networks in classification of multisource and very-high-dimensional remote sensing data. Int. J. Remote Sens. 1993, 14, 2883–2903. [Google Scholar] [CrossRef]
  18. Yang, H. A back-propagation neural network for mineralogical mapping from AVIRIS data. Int. J. Remote Sens. 1999, 20, 97–110. [Google Scholar] [CrossRef]
  19. Gómez-Chova, L.; Camps-Valls, G.; Munoz-Mari, J.; Calpe, J. Semisupervised image classification with Laplacian support vector machines. IEEE Geosci. Remote Sens. Lett. 2008, 5, 336–340. [Google Scholar] [CrossRef]
  20. Bruzzone, L.; Chi, M.; Marconcini, M. A novel transductive SVM for semisupervised classification of remote-sensing images. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3363–3373. [Google Scholar] [CrossRef]
  21. Ma, L.; Crawford, M.M.; Yang, X.; Guo, Y. Local-manifold-learning-based graph construction for semisupervised hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2832–2844. [Google Scholar] [CrossRef]
  22. Ma, L.; Ma, A.; Ju, C.; Li, X. Graph-based semi-supervised learning for spectral-spatial hyperspectral image classification. Pattern Recognit. Lett. 2016, 83, 133–142. [Google Scholar] [CrossRef]
  23. Wang, Z.; Nasrabadi, N.M.; Huang, T.S. Semisupervised hyperspectral classification using task-driven dictionary learning with Laplacian regularization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1161–1173. [Google Scholar] [CrossRef]
  24. Kotsia, I.; Guo, W.; Patras, I. Higher rank support tensor machines for visual recognition. Pattern Recognit. 2012, 45, 4192–4203. [Google Scholar] [CrossRef]
  25. Zhou, H.; Li, L.; Zhu, H. Tensor regression with applications in neuroimaging data analysis. J. Am. Stat. Assoc. 2013, 108, 540–552. [Google Scholar] [CrossRef]
  26. Makantasis, K.; Doulamis, A.D.; Doulamis, N.D.; Nikitakis, A. Tensor-based classification models for hyperspectral data analysis. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6884–6898. [Google Scholar] [CrossRef]
  27. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  28. Zhang, L.; Zhang, L.; Tao, D.; Huang, X. On combining multiple features for hyperspectral remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2012, 50, 879–893. [Google Scholar] [CrossRef]
  29. Shi, M.; Healey, G. Using multiband correlation models for the invariant recognition of 3-D hyperspectral textures. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1201–1209. [Google Scholar]
  30. Zhang, L.; Zhang, L.; Du, B. Deep learning for remote sensing data: A technical tutorial on the state of the art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
  31. Wang, Q.; Liu, S.; Chanussot, J.; Li, X. Scene classification with recurrent attention of VHR remote sensing images. IEEE Trans. Geosci. Remote Sens. 2018, 1–13. [Google Scholar] [CrossRef]
  32. Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
  33. Chen, Y.; Zhao, X.; Jia, X. Spectral–spatial classification of hyperspectral data based on deep belief network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2381–2392. [Google Scholar] [CrossRef]
  34. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  35. Hu, W.; Huang, Y.; Wei, L.; Zhang, F.; Li, H. Deep convolutional neural networks for hyperspectral image classification. J. Sens. 2015, 2015, 1–12. [Google Scholar] [CrossRef]
  36. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  37. Li, Y.; Zhang, H.; Shen, Q. Spectral–spatial classification of hyperspectral imagery with 3D convolutional neural network. Remote Sens. 2017, 9, 67. [Google Scholar] [CrossRef]
  38. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  39. Mikolov, T.; Karafiát, M.; Burget, L.; Černockỳ, J.; Khudanpur, S. Recurrent neural network based language model. In Proceedings of the Eleventh Annual Conference of the International Speech Communication Association, Chiba, Japan, 26–30 September 2010. [Google Scholar]
  40. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  41. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  42. Ienco, D.; Gaetano, R.; Dupaquier, C.; Maurel, P. Land cover classification via multitemporal spatial data by deep recurrent neural networks. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1685–1689. [Google Scholar] [CrossRef]
  43. Sharma, A.; Liu, X.; Yang, X. Land cover classification from multi-temporal, multi-spectral remotely sensed imagery using patch-based recurrent neural networks. Neural Netw. 2018, 105, 346–355. [Google Scholar] [CrossRef] [Green Version]
  44. Mou, L.; Ghamisi, P.; Zhu, X.X. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3639–3655. [Google Scholar] [CrossRef]
  45. Wu, H.; Prasad, S. Convolutional recurrent neural networks for hyperspectral data classification. Remote Sens. 2017, 9, 298. [Google Scholar] [CrossRef]
  46. Shi, C.; Pun, C.M. Multi-scale hierarchical recurrent neural networks for hyperspectral image classification. Neurocomputing 2018, 294, 82–93. [Google Scholar] [CrossRef]
  47. Romera-Paredes, B.; Torr, P.H.S. Recurrent instance segmentation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 312–329. [Google Scholar]
  48. Zhong, Y.; Zhao, J.; Zhang, L. A hybrid object-oriented conditional random field classification framework for high spatial resolution remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7023–7037. [Google Scholar] [CrossRef]
  49. Yin, W.; Kann, K.; Yu, M.; Schütze, H. Comparative study of cnn and rnn for natural language processing. arXiv, 2017; arXiv:1702.01923. [Google Scholar]
  50. Fan, F.; Deng, Y. Enhancing endmember selection in multiple endmember spectral mixture analysis (MESMA) for urban impervious surface area mapping using spectral angle and spectral distance parameters. Int. J. Appl. Earth Obs. Geoinf. 2014, 33, 290–301. [Google Scholar] [CrossRef]
  51. Roweis, S.T.; Saul, L.K. Nonlinear dimensionality reduction by locally linear embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
  52. Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
  53. Asner, G.P.; Heidebrecht, K.B. Spectral unmixing of vegetation, soil and dry carbon cover in arid regions: Comparing multispectral and hyperspectral observations. Int. J. Remote Sens. 2002, 23, 3939–3958. [Google Scholar] [CrossRef]
  54. Tobler, W.R. A computer movie simulating urban growth in the Detroit region. Econ. Geogr. 1970, 46, 234–240. [Google Scholar] [CrossRef]
  55. Pu, H.; Chen, Z.; Wang, B.; Jiang, G.M. A novel spatial–spectral similarity measure for dimensionality reduction and classification of hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2014, 52, 7008–7022. [Google Scholar]
  56. Huttenlocher, D.P.; Klanderman, G.A.; Rucklidge, W.J. Comparing images using the Hausdorff distance. IEEE Trans. Pattern Anal. Mach. Intell. 1993, 15, 850–863. [Google Scholar] [CrossRef]
  57. Congalton, R.G. A review of assessing the accuracy of classifications of remotely sensed data. Remote Sens. Environ. 1991, 37, 35–46. [Google Scholar] [CrossRef]
Figure 1. Architecture of the proposed methods.
Figure 1. Architecture of the proposed methods.
Remotesensing 11 00194 g001
Figure 2. The basic structure of an unrolled long short-term memory LSTM unit.
Figure 2. The basic structure of an unrolled long short-term memory LSTM unit.
Remotesensing 11 00194 g002
Figure 3. False-color image composites and their corresponding ground-reference data. (a) false-color composite of Pavia University image bands (R: band 55, G: band 33, and B: band 13); (b) ground-reference data (with class legend) for the Pavia University image; (c) false-color composite of Salinas image bands (R: band 57, G: band 27, and G: band 17); (d) ground-reference data (with class legend) for the Salinas image.
Figure 3. False-color image composites and their corresponding ground-reference data. (a) false-color composite of Pavia University image bands (R: band 55, G: band 33, and B: band 13); (b) ground-reference data (with class legend) for the Pavia University image; (c) false-color composite of Salinas image bands (R: band 57, G: band 27, and G: band 17); (d) ground-reference data (with class legend) for the Salinas image.
Remotesensing 11 00194 g003
Figure 4. Classification maps for the Pavia University image from the fifth trial: (a) ground-reference map; (b) SVM, with OA = 80.04%; (c) 1DCNN, with OA = 78.32%; (d) 1DLSTM, with OA = 83.72%; (e) LSTM_PM_EU, with OA = 83.81%; (f) LSTM_PM_SAM, with OA = 86.78%; (g) LSTM_BM_EU, with OA = 93.18%; and (h) LSTM_BM_SAM, with OA = 96.01%. The red-rectangle and red-circle annotations represent sample areas of interest, discussed in the text.
Figure 4. Classification maps for the Pavia University image from the fifth trial: (a) ground-reference map; (b) SVM, with OA = 80.04%; (c) 1DCNN, with OA = 78.32%; (d) 1DLSTM, with OA = 83.72%; (e) LSTM_PM_EU, with OA = 83.81%; (f) LSTM_PM_SAM, with OA = 86.78%; (g) LSTM_BM_EU, with OA = 93.18%; and (h) LSTM_BM_SAM, with OA = 96.01%. The red-rectangle and red-circle annotations represent sample areas of interest, discussed in the text.
Remotesensing 11 00194 g004
Figure 5. Classification maps of Salinas image from the fifth trial: (a) ground-reference map; (b) SVM, with OA = 83.38%; (c) 1DCNN, with OA = 87.00%; (d) 1DLSTM, with OA = 86.85%; (e) LSTM_PM_EU, with OA = 86.21%; (f) LSTM_PM_SAM, with OA = 87.13%; (g) LSTM_BM_EU, with OA = 90.02%; and (h) LSTM_BM_SAM, with OA = 90.72%. The red-circle and red-rectangle annotations represent sample areas of interest, discussed in the text.
Figure 5. Classification maps of Salinas image from the fifth trial: (a) ground-reference map; (b) SVM, with OA = 83.38%; (c) 1DCNN, with OA = 87.00%; (d) 1DLSTM, with OA = 86.85%; (e) LSTM_PM_EU, with OA = 86.21%; (f) LSTM_PM_SAM, with OA = 87.13%; (g) LSTM_BM_EU, with OA = 90.02%; and (h) LSTM_BM_SAM, with OA = 90.72%. The red-circle and red-rectangle annotations represent sample areas of interest, discussed in the text.
Remotesensing 11 00194 g005
Figure 6. Analysis of different sequential feature lengths. (a) OAs based on the Pavia University image data; and (b) OAs based on the Salinas image data.
Figure 6. Analysis of different sequential feature lengths. (a) OAs based on the Pavia University image data; and (b) OAs based on the Salinas image data.
Remotesensing 11 00194 g006
Figure 7. Analysis of different window sizes. (a) OAs based on the Pavia University image data; and (b) OAs based on the Salinas image data.
Figure 7. Analysis of different window sizes. (a) OAs based on the Pavia University image data; and (b) OAs based on the Salinas image data.
Remotesensing 11 00194 g007
Table 1. Class codes for Pavia University and Salinas images.
Table 1. Class codes for Pavia University and Salinas images.
Pavia University ImageSalinas image
Class No.NameClass No.Name
1Asphalt1Brocoli_green_weeds_1
2Meadow2Brocoli_green_weeds_2
3Gravel3Fallow
4Trees4Fallow_rough_plow
5Painted Metal Sheets5Fallow_smooth
6Bare Soil6Stubble
7Bitumen7Celery
8Self-Blocking Bricks8Grapes_untrained
9Shadows9Soil_vinyard_develop
10Corn_senesced_green_weeds
11Lettuce_romaine_4wk
12Lettuce_romaine_5wk
13Lettuce_romaine_6wk
14Lettuce_romaine_7wk
15Vinyard_untrained
16Vinyard_vertical_trellis
Table 2. Parameter settings for 1D-CNN and 1D-LSTM, where the convolutional layer is represented as “Conv(number of kernels)-(kernel size)”, maxpooling layer is performed as “Maxpooling-(kernel size)”, and LSTM layer is denoted as “LSTM-(kernel size)”.
Table 2. Parameter settings for 1D-CNN and 1D-LSTM, where the convolutional layer is represented as “Conv(number of kernels)-(kernel size)”, maxpooling layer is performed as “Maxpooling-(kernel size)”, and LSTM layer is denoted as “LSTM-(kernel size)”.
Pavia University ImageSalinas Image
1DCNN1DLSTM1DCNN1DLSTM
Conv(10)-8LSTM-32Conv(8)-12LSTM-32
Maxpooling-2LSTM-64Maxpooling-2LSTM-64
Conv(10)-8LSTM-128Conv(8)-12LSTM-128
Maxpooling-2 Maxpooling-2
FC layer-9FC layer-16
Table 3. Comparison of different classification accuracy results for the Pavia University image (%), where LSTM_PEU equals LSTM_PM_EU, LSTM_PSAM equals LSTM_PM_SAM, LSTM_BEU equals LSTM_BM_EU, and LSTM_BSAM equals LSTM_BM_SAM.
Table 3. Comparison of different classification accuracy results for the Pavia University image (%), where LSTM_PEU equals LSTM_PM_EU, LSTM_PSAM equals LSTM_PM_SAM, LSTM_BEU equals LSTM_BM_EU, and LSTM_BSAM equals LSTM_BM_SAM.
Class No.SVM1DCNN1DLSTMLSTM_PEULSTM_PSAMLSTM_BEULSTM_BSAM
197.52 ± 0.2196.53 ± 0.5795.76 ± 0.6585.51 ± 3.0694.06 ± 0.7298.76 ± 0.4498.78 ± 0.39
295.77 ± 0.3094.78 ± 1.5594.13 ± 0.9394.11 ± 1.4096.01 ± 1.0899.08 ± 0.2598.84 ± 0.29
365.57 ± 3.4168.93 ± 3.6564.29 ± 3.0270.41 ± 2.2766.61 ± 3.6890.22 ± 2.2889.22 ± 2.63
471.27 ± 7.3876.62 ± 7.9275.71 ± 4.9562.69 ± 6.2881.58 ± 4.7792.97 ± 2.1994.70 ± 1.71
595.50 ± 1.5598.50 ± 0.8697.73 ± 1.3198.51 ± 0.6795.57 ± 2.8098.99 ± 1.1899.42 ± 0.64
659.51 ± 6.7463.62 ± 11.0361.12 ± 7.6269.71 ± 8.1864.39 ± 8.2388.65 ± 4.9390.65 ± 3.92
752.10 ± 0.9366.87 ± 4.8665.74 ± 3.6263.76 ± 4.4556.69 ± 4.4289.72 ± 4.2588.47 ± 6.60
884.27 ± 1.1383.54 ± 1.7381.48 ± 1.7082.37 ± 1.5980.50 ±  1.0193.32 ± 2.4693.27 ± 1.28
999.92 ± 0.1199.65 ± 0.3199.72 ± 0.3593.74 ± 2.0595.99 ± 5.0699.09 ± 0.6298.49 ± 1.75
OA82.12 ± 1.7984.45 ± 3.0183.41 ± 2.6682.70 ± 1.7384.56 ± 2.4195.96 ± 1.0196.20 ± 0.57
AA80.16 ± 0.9683.23 ± 1.7181.74 ± 1.4780.09 ± 1.1981.27 ± 1.7294.53 ± 0.9894.65 ± 0.64
Kappa76.98 ± 2.1279.79 ± 3.5378.42 ± 3.1377.40 ± 2.0579.86 ± 2.8994.60 ± 1.3194.91 ± 0.75
Table 4. Comparison of different classification results for the Salinas image (%), where LSTM_PEU equals LSTM_PM_EU, LSTM_PSAM equals LSTM_PM_SAM, LSTM_BEU equals LSTM_BM_EU, and LSTM_BSAM equals LSTM_BM_SAM.
Table 4. Comparison of different classification results for the Salinas image (%), where LSTM_PEU equals LSTM_PM_EU, LSTM_PSAM equals LSTM_PM_SAM, LSTM_BEU equals LSTM_BM_EU, and LSTM_BSAM equals LSTM_BM_SAM.
Class No.SVM1DCNN1DLSTMLSTM_PEULSTM_PSAMLSTM_BEULSTM_BSAM
196.84 ± 1.1899.12 ± 1.5294.18 ± 10.0498.61 ± 3.9399.40 ± 0.8995.78 ± 11.3099.86 ± 0.22
298.79 ± 0.1398.79 ± 0.6998.69 ± 0.7898.75 ± 1.0899.39 ± 0.2899.60 ± 0.1699.34 ± 0.51
385.11 ± 1.3895.53 ± 1.3288.68 ± 7.3292.89 ± 3.6895.70 ± 1.0796.11 ± 1.2596.24 ± 1.10
497.44 ± 0.1897.94 ± 0.6797.20 ± 1.0298.33 ± 0.7498.45 ± 0.6898.66 ± 0.9597.90 ± 1.70
595.03 ± 0.8597.20 ± 2.7698.07 ± 1.6596.26 ± 7.5097.63 ± 1.5798.75 ± 1.6598.83 ± 0.68
699.79 ± 0.1199.77 ± 0.2198.98 ± 1.1699.64 ± 0.4499.45 ± 0.8799.69 ± 0.3699.71 ± 0.33
798.63 ± 0.4499.35 ± 0.6298.81 ± 0.9199.37 ± 0.5399.11 ± 0.7699.40 ± 0.4399.38 ± 0.51
876.70 ± 1.3183.99 ± 4.0276.24 ± 5.9076.97 ± 2.9176.50 ± 3.4883.04 ± 1.3785.66 ± 3.00
999.12 ± 0.0499.02 ± 0.2998.03 ± 1.4698.78 ± 0.2898.66 ± 0.3399.20 ± 0.4199.43 ± 0.27
1081.91 ± 1.5885.89 ± 2.0984.90 ± 1.9485.94 ± 4.1988.65 ± 1.2094.16 ± 1.5891.00 ± 2.13
1169.51 ± 1.0082.23 ± 6.7181.23 ± 13.7976.81 ± 8.0683.24 ± 4.3386.00 ± 3.2783.49 ± 4.40
1293.33 ± 0.3496.82 ± 1.0587.58 ± 9.5797.09 ± 1.3397.26 ± 1.0398.31 ± 1.3698.26 ± 1.67
1392.67 ± 0.6594.15 ± 2.6290.12 ± 4.0596.07 ± 2.2395.27 ± 2.2795.89 ± 2.7897.33 ± 1.98
1489.68 ± 2.1989.80 ± 3.7884.77 ± 12.6187.66 ± 5.9588.47 ± 5.3492.50 ± 3.9590.75 ± 3.77
1556.78 ± 2.3959.34 ± 12.3856.88 ± 8.1360.79 ± 6.7263.62 ± 5.5867.33 ± 2.8269.85 ± 3.21
1695.59 ± 1.1898.06 ± 0.4493.84 ± 1.5996.53 ± 2.1496.63 ± 1.1298.03 ± 1.0796.19 ± 3.79
OA84.75 ± 0.6285.99 ± 4.1484.07 ± 2.9086.35 ± 2.0787.53 ± 0.9589.90 ± 0.4390.63 ± 0.61
AA89.18 ± 0.2392.31 ± 0.9689.26 ± 3.0291.28 ± 1.4092.34 ± 0.5693.90 ± 0.7193.95 ± 0.55
Kappa76.98 ± 2.1284.45 ± 4.4982.26 ± 3.1984.78 ± 2.2686.07 ± 1.0588.72 ± 0.4889.55 ± 0.68
Table 5. Average training time (min) of LSTM models.
Table 5. Average training time (min) of LSTM models.
LSTM ParameterPavia University Image
Sequential Feature LengthLSTM_PM_EULSTM_PM_SAMLSTM_BM_EULSTM_BM_SAM
1030.8834.0129.1132.00
2051.1453.2549.9253.73
3083.3987.4377.5281.63
40112.49115.03105.31110.75
50145.29144.98132.31142.07
Salinas Image
Sequential Feature LengthLSTM_PM_EULSTM_PM_SAMLSTM_BM_EULSTM_BM_SAM
1029.0834.1928.4728.00
2046.2749.0244.9647.64
3076.9485.6175.8375.41
40110.57117.78101.10102.07
50124.06147.91128.13128.90

Share and Cite

MDPI and ACS Style

Ma, A.; Filippi, A.M.; Wang, Z.; Yin, Z. Hyperspectral Image Classification Using Similarity Measurements-Based Deep Recurrent Neural Networks. Remote Sens. 2019, 11, 194. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11020194

AMA Style

Ma A, Filippi AM, Wang Z, Yin Z. Hyperspectral Image Classification Using Similarity Measurements-Based Deep Recurrent Neural Networks. Remote Sensing. 2019; 11(2):194. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11020194

Chicago/Turabian Style

Ma, Andong, Anthony M. Filippi, Zhangyang Wang, and Zhengcong Yin. 2019. "Hyperspectral Image Classification Using Similarity Measurements-Based Deep Recurrent Neural Networks" Remote Sensing 11, no. 2: 194. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11020194

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop