Next Article in Journal
A Novel Approach to Modelling Mangrove Phenology from Satellite Images: A Case Study from Northern Australia
Next Article in Special Issue
Multilevel Structure Extraction-Based Multi-Sensor Data Fusion
Previous Article in Journal
Random Sample Fitting Method to Determine the Planetary Boundary Layer Height Using Satellite-Based Lidar Backscatter Profiles
Previous Article in Special Issue
A Hybrid Attention-Aware Fusion Network (HAFNet) for Building Extraction from High-Resolution Imagery and LiDAR Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Fusion Using a Multi-Sensor Sparse-Based Clustering Algorithm

1
Helmholtz-Zentrum Dresden-Rossendorf, Helmholtz Institute Freiberg for Resource Technology, 09599 Freiberg, Germany
2
Imec-Vision Lab, Department of Physics, University of Antwerp, 2000 Antwerp, Belgium
*
Author to whom correspondence should be addressed.
Remote Sens. 2020, 12(23), 4007; https://0-doi-org.brum.beds.ac.uk/10.3390/rs12234007
Submission received: 29 October 2020 / Revised: 27 November 2020 / Accepted: 3 December 2020 / Published: 7 December 2020
(This article belongs to the Special Issue Advanced Multisensor Image Analysis Techniques for Land-Cover Mapping)

Abstract

:
The increasing amount of information acquired by imaging sensors in Earth Sciences results in the availability of a multitude of complementary data (e.g., spectral, spatial, elevation) for monitoring of the Earth’s surface. Many studies were devoted to investigating the usage of multi-sensor data sets in the performance of supervised learning-based approaches at various tasks (i.e., classification and regression) while unsupervised learning-based approaches have received less attention. In this paper, we propose a new approach to fuse multiple data sets from imaging sensors using a multi-sensor sparse-based clustering algorithm (Multi-SSC). A technique for the extraction of spatial features (i.e., morphological profiles (MPs) and invariant attribute profiles (IAPs)) is applied to high spatial-resolution data to derive the spatial and contextual information. This information is then fused with spectrally rich data such as multi- or hyperspectral data. In order to fuse multi-sensor data sets a hierarchical sparse subspace clustering approach is employed. More specifically, a lasso-based binary algorithm is used to fuse the spectral and spatial information prior to automatic clustering. The proposed framework ensures that the generated clustering map is smooth and preserves the spatial structures of the scene. In order to evaluate the generalization capability of the proposed approach, we investigate its performance not only on diverse scenes but also on different sensors and data types. The first two data sets are geological data sets, which consist of hyperspectral and RGB data. The third data set is the well-known benchmark Trento data set, including hyperspectral and LiDAR data. Experimental results indicate that this novel multi-sensor clustering algorithm can provide an accurate clustering map compared to the state-of-the-art sparse subspace-based clustering algorithms.

Graphical Abstract

1. Introduction

The recent advances in imaging technology have resulted in the acquisition of high spectral-spatial resolution imaging data [1]. Among the designed imaging sensors, hyperspectral cameras have become the main source of fine spectral-resolution data in hundreds of narrow spectral bands [2]. The spectral channels (bands) of hyperspectral imaging data (HSI) cover the electromagnetic spectrum ranging from the visible to the near-infrared (VNIR, 0.4–1 μ m) and shortwave infrared (SWIR, 1–2.5 μ m). An HSI can be employed to identify and discriminate different materials and biological targets [2]. The rich spectral information derived from an HSI has been widely used in remote sensing applications such as land-cover- [3], urban- [4] and mineral-mapping [5,6], or the estimation of above ground biomass [7]. For instance, in the mineral exploration stage, information that is obtained by an HSI can be used as a useful support for geologists to distinguish and track mineralogical features of interest [8].
The analysis of HSI data can be challenging due to (1) their high dimensionality and the often limited availability of representative samples (also known as training samples ), which results in the so-called curse of dimensionality [9] and (2) the highly mixed nature of pixels [10]. In order to overcome these challenges, different machine learning algorithms have been developed and proposed in the last decades [11]. Supervised learning algorithms have become popular for HSI analysis because of their empirical success and adjustability to solve classification and regression problems [12,13]. However, these algorithms strongly rely on training samples, which still remain challenging to obtain.
This issue has been tackled by developing unsupervised algorithms (also known as clustering algorithms) [14]. Clustering algorithms use unlabeled data to group the data into meaningful clusters. Conventional clustering algorithms (e.g., K-means [15], Fuzzy C-means [16]) distribute the data around several cluster centers, the so-called centroids. By using different distance measures, data points are assigned to the cluster with the closest centroid [15]. These clustering algorithms are iterative algorithms and strongly depend on the random initialization of the centroids [17]. Moreover, conventional clustering algorithms are pixel-based, and do not include any spatial information. Additionally, they are only applicable to single-sensor data.
In the past decades, HSIs have been extensively used to detect and characterize different materials, as they provide rich spectral information. At the same time, RGB sensors can generate very detailed high spatial resolution images, while light detection and ranging sensors (LiDAR) can yield geometrical and elevation data [18,19]. All the above-mentioned multi-sensor data sets lead to complementary information (i.e., spectral, spatial, elevation), and the fusion of these data sets has received much attention. In many studies (to name a few [4,5,19,20,21,22,23,24]), the integration of spectral, spatial and elevation information, extracted from multi-sensors has led to significant improvements in the performance of supervised learning-based approaches. In [19], a low-rank subspace-based fusion technique (SubFus) is proposed to incorporate contextual features from multi-sensor data for classification. In [5], the authors proposed to fuse HSI and fine spatial-resolution RGB images using orthogonal total variation component analysis (OTVCA) as a spectral-spatial feature extraction technique. In [21], WorldView 2/3 and LiDAR images were integrated using a deep learning algorithm, the so-called dense convolutional network to map individual tree species in urban areas. A comprehensive review on the integration of multi-sensor and multi-temporal data sets is given in [4].
It has been demonstrated in different studies that high dimensional data can lie in a union of lower dimensional subspaces [14,17,25,26,27,28]. In this way, data points can be grouped into meaningful clusters using subspace information rather than using the complete high dimensional data, and thus reducing the complexity and computational cost [17]. Among the subspace-based clustering algorithms, sparse representation-based algorithms have received much attention in many fields (e.g., feature fusion [29] and image segmentation [14,17,30,31]). The main motivation to use sparsity-based clustering is that the sparse subspace-based clustering algorithm (SSC) was proven to be robust against noise [32]. In addition, the complexity of the optimization with sparse representations is lower than with other representations such as low-rank representations. The sparse subspace clustering method (SSC) is a well-known clustering algorithm, which benefits from the self-expressiveness property. This property states that data points of an HSI can be written as a linear combination of other data points from the same subspace [17]. To solve the SSC optimization problem, the alternative direction method of multipliers (ADMM) is employed. Nevertheless, SSC has some drawbacks: (1) it uses the entire data set for the clustering process, which makes it hard to handle large data sets (this is known as the scalability issue), (2) it only uses spectral information and does not account for the spatial continuity between the data points [14].
During the last decades, several studies have been conducted to cope with the scalability issue. In [33], the authors proposed a landmark-based spectral clustering method (LSC) to handle large data sets. In LSC, spectral clustering is applied on a few representative points (also known as landmark points), which are selected from the data by clustering or random sampling. In [34], a sparse subspace clustering method by using orthogonal matching pursuit (SSC-OMP) was proposed. In [14], an exemplar-based subspace clustering method (ESC) was proposed, which searches for a subset of representative samples, the so-called exemplars that can describe the entire data. Moreover, such an approach allows ESC to reduce the computational cost and to analyze larger data sets. However, the random initialization in search for the exemplars of a data set, results in a less robust performance of ESC. We recently proposed a hierarchical sparse subspace-based clustering (HESSC), as a fast and robust HSI clustering algorithm [28]. In HESSC, the entire data set is explained in different levels of detail using a hierarchical structure, using a lasso-based binary tree and a criterion to split nodes and automatically determine the number of clusters. Nonetheless, all the introduced advanced sparse subspace-based clustering algorithms merely utilize spectral information.
On the other hand, several methods were developed to incorporate spatial information in the sparse subspace-based clustering procedure. In [35], a spatial subspace-based clustering algorithm (SpatSC) was proposed to employ spatial constraints in favor of utilizing information derived from neighboring pixels that have similar features in mineralogical and geological structures. In [26,36], spatial information is exploited by applying a mean or median filter on the computed sparse coefficient matrix in the upgrade phase of ADMM. The above-mentioned algorithms use the entire HSI for clustering, and do not cope with the scalability issue. Additionally, all the mentioned spectral-spatial clustering algorithms use information derived from single-sensor data.
In recent years, several morphological-based algorithms were developed that exploit spatial information for image processing. In [37], for the first time, it was proposed to use morphological profiles (MPs), using opening and closing by reconstruction operators with different structuring elements (SEs) to segment high resolution satellite images. As a follow-up, in [38], extended MPs (EMPs) were introduced as the MPs, which are extracted from the first principle components of an HSI. MPs and EMPs are powerful tools to extract spatial information, but are suffering from a few limitations: (1) the SEs have fixed shapes, (2) they are not suitable to map all the different characteristics of spatial information. In [39], morphological attribute profiles (APs) were proposed to extract structural information of very high resolution images. In APs, morphological attribute filters (i.e., attribute thinning and thickening) are applied on connected components of a gray-scale image to extract different spatial characteristics of a scene. Therefore, APs can be considered as a generalization of MPs that are better able to capture region-based attributes. In order to handle HSIs with APs, extended morphological APs (EMAPs) were proposed [40]. Additionally, in [41,42], different feature extraction techniques (e.g., kernel PCA, multi-linear PCA, and the like) were investigated on the performances of EMPs and EMAPs.
Nevertheless, APs may suffer from an under-segmentation problem, as it may group different regions belonging to several objects. To overcome this problem, invariant attribute profiles (IAPs) were proposed recently [43]. IAPs are built in two phases, the first phase is a spatial invariant feature extraction, where isotropic filtering (by convolutional kernels) is employed to extract the convolutional features of the HSI. This process is followed by a simple linear iterative clustering to spatially aggregate the filtered features. In the second phase, continuous APs are obtained by modeling the variant behaviors of pixels of the HSI. The aforementioned procedure is applied in a Fourier polar coordinate system to avoid discretization artifacts and capture frequency features [44].
In this paper, we propose a spectral-spatial multi-sensor-based clustering algorithm. The proposed algorithm extracts information from both HSI and fine spatial-resolution images. Spectral information is extracted from the HSI, while spatial features (i.e., MPs and IAPs) contain the spatial and elevation information from RGB and LiDAR data. In order to fuse and cluster the obtained information from the different sensors, the concept of HESSC is adopted, in which a fusion between the spectral and spatial features is accomplished at each hierarchical level of the binary tree structure.
The following items summarize the proposed contribution of this paper:
1.
We propose a novel multi-sensor sparse-based clustering algorithm that describes the data at different levels of detail.
2.
To the best of our knowledge, this is the first attempt to incorporate spatial information in the form of morphological-based profiles extracted from multi-sensor data sets in a hierarchical sparse-based clustering algorithm.
3.
In the proposed algorithm, both spectral and spatial features equally contribute at each level of the tree.
4.
The proposed algorithm is able to cluster large-scale data sets.
5.
The proposed algorithm can be adapted to different ancillary remote sensing data sets (e.g., RGB, multi-spectral images, HSI, LiDAR, synthetic aperture radar).

2. Methodology

In this section, we provide a brief overview of the proposed methodology. The workflow of the proposed algorithm is presented in Figure 1. As can be observed from the figure, spatial features are extracted from high-spatial resolution images (i.e., RGB, LiDAR). The process is then followed by establishing a hierarchical (tree) structure to describe the data at various levels of detail. In order to build the tree structure and fuse the spectral and spatial features at the level of the nodes, a lasso-based binary clustering was employed. Finally, if an end-node cannot be divided further, all data points corresponding to this end-node are assigned to one cluster.

2.1. Spatial Feature Extraction

We start to describe the spatial features that are obtained from the auxiliary high spatial-resolution image. As the spatial features, we propose to employ MPs and IAPs, the concepts of which will be described in detail in the following paragraphs.

2.1.1. Morphological Profiles

In [37], extracted MPs rely on two mathematical morphology operators (i.e., opening and closing). These operators are applied on an image with a series of structuring elements (SEs), that can have different shapes (e.g., disc, square, cube) with a user-defined size r . With morphological opening, an eroded image is dilated to capture the bright structures in a scene, while with morphological closing, a dilated image is eroded to filter out the dark structures in the scene. Additionally, to preserve the original image structure, a reconstruction filter is applied after the aforementioned morphological operators [41].
Let us assume that X = [ x 1 , x 2 , . . . , x N ] T R b × N expresses a high spatial-resolution image, where b is the number of spectral channels (e.g., 3 for an RGB image), and N represents the number of pixels in the image. In addition, the location of each pixel in X is denoted by ( x , y ) in the Cartesian coordinates. Let γ S E ( X ) and δ S E ( X ) denote the opening and closing by reconstruction operators on X with a set of SEs, respectively. In this study, only disk-shaped SEs with variable size are used. MPs are then computed as:
M P γ = { M P γ r ( X ) = γ r ( X ) , r [ 0 , n ] } M P δ = { M P δ r ( X ) = δ r ( X ) , r [ 0 , n ] } S . t . γ 0 ( X ) = δ 0 ( X ) = X
where r is the radius of a disk-shaped SE, and M P γ and M P δ are the extracted opening and closing MPs by reconstruction [38]; n is a non-negative integer. For more information, we refer to [37,38]. MPs have been widely applied to extract spatial information in remote sensing applications, although they cannot provide all characteristics of the spatial information [11].

2.1.2. Invariant Attribute Profiles

In [43], invariant attribute profiles (IAPs) were defined as the concatenation of two different types of invariant features, i.e., spatial invariant features (SIFs) and frequency invariant features (FIFs).
In order to extract the SIFs, initially, robust convolutional features F R C F are obtained by isotropic filtering of X . For this, each band of the image is convolved with a convolutional kernel in order to aggregate the local spatial information. Then, a simple linear iterative clustering (SLIC), a well-known super-pixel segmentation approach, is applied on X [45]. Eventually, the extracted convolutional features F R C F are spatially aggregated by the super-pixels. For each pixel i, the SIFs are defined as:
F S I F s i = 1 N q j = 1 N q F R C F j , j ϕ i , q
where ϕ i , q is the pixel set of the q-th super-pixel, containing the i-th targeted pixel and a total of N q pixels.
To derive the frequency invariant features (FIFs), first a polarized Fourier transformation is applied on X . In this way, rotation invariance is obtained by accessing the angular information. From this transformation, three groups of rotation-invariant features are calculated: F m 1 ( x , y ) are magnitude features corresponding to m different Fourier orders; F m 2 ( x , y ) are absolute rotation-invariant features, obtained by removing the phase information; F m 3 ( x , y ) are relative rotation-invariant features, that compensate for the loss of rich phase information. The FIFs are obtained by concatenating these rotation-invariant features and aggregating them into region-based features by using isotropic convolutional filters.
Finally, IAPs are produced by stacking the SIFs and the FIFs: F I A P s = [ F S I F s F F I F s ] . More detailed information on the generation of the IAPs can be found in [43].

2.2. Sparse Subspace Clustering (SSC)

Let Y denote the input HSI with size D × N , where D is the number of spectral bands. y i represents the ith column vector of Y , where i { 1 , 2 , . . . , N } is the pixel index in the HSI, and y j i is the element in row j and column i and denotes band j of that pixel.
The main assumption in sparse subspace clustering methods is that each data point can be described as a linear combination of a dictionary of data points from the same subspace and that all pixels are drawn from a union of L linear subspaces { S l } l = 1 L of dimension { d l } l = 1 L [17]. Therefore, Y = Y = Y 1 Y 2 Y 3 . . . Y L , where each Y l R D × N l is a sub-matrix of Y , containing N l pixels from subspace S l . Remark that the number of pixels of each subspace should be larger than the dimension of the subspace ( N l > d l ). In most subspace-based clustering algorithms, the number and the dimensions of the subspaces need to be known a priori.
SSC addresses the aforementioned issue by employing the data itself as a dictionary. Accordingly, the following sparse optimization problem is obtained:
arg min C | | C | | 1 + λ 2 | | Y Y C | | F 2 S . t . , d i a g ( C ) = 0 ,
where C = [ c 1 , c 2 , . . . , c N ] R N × N is the sparse coefficient matrix derived from Y . | | . | | 1 is the 1 -norm: | | C | | 1 = Σ i = 1 N ( Σ j = 1 N | c j i | ) , and | | C | | F = Σ i = 1 N ( Σ j = 1 N | c j i | 2 ) 1 2 represents the Frobenius-norm [17]. λ is a trade off parameter between the sparsity and fidelity terms. The constraint d i a g ( C ) = 0 is required to avoid a point being represented as a linear combination of itself. The alternating direction method of multipliers (ADMM) solver is employed to solve the sparse optimization problem of Equation (3) [46]. Consequently, the non-zero elements ( c j i ) of C identify all data points j that are from the same subspace as i. From C , a symmetrical similarity matrix W = | C | + | C | T is obtained, to assure that all pixels from the same subspace are connected to each other.
As the final step, to reduce the dimensionality of the problem, a spectral clustering algorithm [47] is applied on W . For this, the normalized Laplacian matrix is computed: L : = D W , where D R N × N is a diagonal matrix, with d i i = j = 1 N w i j . Then, singular-value decomposition (SVD) is applied on L , to obtain the eigenvectors V R N × A , corresponding to the A smallest non-zero eigenvalues of L . Finally, clustering (we used K-means clustering) is applied on V to generate the clustering map [17,47].
The main advantage of SSC is that the clustering is applied on the similarity matrix of the sparse coefficient matrix, rather than directly on the data. In this way, the subspace structure of the data is accounted for, and makes the clustering more robust to the initialization compared to the conventional clustering algorithms (e.g., K-means, FCM) [48].
Although SSC performs well to cluster HSIs, it also has shortcomings. SSC has a scalability issue, in which it requires to use all data points to solve the optimization problem in Equation (3), which can be computationally expensive in the case of large data sets. Another shortcoming of SSC is that the algorithm only uses spectral information and does not take into account the spatial context of the HSI.

2.3. Hierarchical Sparse Subspace Clustering (HESSC)

Several sparse subspace-based clustering algorithms have been proposed in order to tackle the scalability issue, [14,27,28,49]. In our earlier work [28], we proposed an automatic hierarchical sparse subspace-based clustering (HESSC) algorithm to analyze HSIs. HESSC is able to describe an HSI at different hierarchical levels in a robust and automatic manner by employing a binary tree structure.
Starting from all data points in the HSI, at each level, nodes are split into two groups. Let us denote Y l i as the collection of all the data points in one of the parental nodes at level l i . Then, Y l i = r = 1 2 Y r , l i + 1 , where Y r , l i + 1 denote the collection of all data points in each of the two child nodes at level i + 1 . The criterion to split the nodes is based on the generation of a binary matrix, obtained by the following sparse optimization problem:
arg min C l i | | C l i | | 1 + λ 2 | | Y l i y j C l i | | F 2 ,
where C l i is the computed sparse coefficients derived from Y l i , and y j is the j-th column vector of Y l i , where j is randomly selected. To solve the above optimization problem, the lasso version of the least angle regression algorithm is used [50]. Subsequently, the computed C l i is normalized, and thresholded (with a user-defined threshold) to generate a binary matrix (BM) (Figure 2). In order to assure the robustness of HESSC, the aforementioned procedure is repeated 100 times, with randomly selected y j , each time generating a binary matrix. Subsequently, the generated BMs are fed to an entropy-based consensus clustering algorithm (ECC) [51] to produce the final binary matrix that eventually splits the current node into two child nodes.
A stopping criterion is applied to decide whether a node should be further split. The stopping criterion is either manual, when the required number of clusters is known prior and pre-defined by the user, or automatic, based on a Singular Value Decomposition of the data [52]. When a node is not split any further, its’ corresponding data points are labeled as one cluster.
Although HESSC is able to cluster an HSI in a robust and fast manner, only spectral information is utilized. In this work, we will extend HESSC to include spatial information that is derived from high spatial-resolution data from the same scene as the HSI.

2.4. Multi-Sensor Spectral-Spatial Sparse-Based Clustering (Multi-SSC)

One way of fusing spectral and spatial information is to extract spatial features from an HSI, and combining (e.g., concatenating) the extracted spatial features with the spectral channels [53]. However, such a single-sensor integration approach might not contain sufficient spatial and contextual information. We therefore propose to explicitly incorporate the spatial information derived from high spatial-resolution data sets (i.e., high spatial-resolution RGB and LiDAR images) [19].
In the proposed Multi-SSC, a hierarchical tree approach, similar to the HESSC algorithm is used. However, at each level of the binary tree, a parental node is split, based on binary matrices, each of which is generated from either spectral or spatial information. To generate a BM from spectral information, the sparse optimization problem of Equation (4) is solved and the obtained sparse coefficient matrix is thresholded. To generate a BM from spatial information, the following sparse optimization problem is defined:
arg min C l i | | C l i | | 1 + λ 2 | | F S F E l i F S F E j C r , l i | | F 2 ,
where F S F E l i denotes the extracted spatial features (either MPs or IAPs) from all data points at the specific parental node of level l i . F S F E j is the j-th column vector of F S F E l i , which has been randomly selected. The obtained sparse coefficient matrix is then thresholded to generate the BM.
To cope with uncertainties of the model [54] that may be introduced by the random selection of both F S F E j and y j , we apply the ECC algorithm [51]. Additionally, to include both spectral and spatial information in a balanced way, an equal number of BMs are generated from the optimizations of Equations (4) and (5). Depending on the specific application, this ratio can be adapted. In this work, we choose to apply each of the optimizations 50 times. The total of 100 BMs are then processed by the ECC algorithm to produce the final binary matrix, which is used for the further division of the current node. The stopping criterion is the same as applied in Section 2.3. The final clustering map is generated by labeling the nodes of the tree, which are not further split.

3. Experiments

3.1. Data Acquisition and Description

In order to evaluate the generalization capability of the proposed method, we take into account two important factors for the selection of data sets: (1) the investigated data sets cover different scene types (i.e., geological and rural areas) and (2) different sensor combinations and types are investigated (i.e., HSI, RGB, and LiDAR). For this reason, we chose the following data sets:
(1)
Czech data : The first data set was acquired near the Litov tailing lake and its adjacent waste heap, situated in the Sokolov district of the Czech Republic. Both HSI and RGB images were acquired. The HSI was acquired by a hyperspectral frame-based camera (0.6 Mp Rikola Hyperspectral Imager), which was deployed on a hexacopter unmanned aerial vehicle (UAV; Aibotix Aibot X6v2) along with a pre-programmed stop-scan-motion flight plan to capture a complete set of HSIs for the subsequent image mosaicking. The RGB image was captured by employing a senseFly S.O.D.A. RGB camera, deployed on a fixed-wing UAV. The spatial resolution of the captured RGB image is 1.5 cm. It is downsampled to the size of the HSI data ( 1066 × 909 pixels), which has a spatial resolution of 3.3 cm and is composed of 50 spectral bands ranging from 0.50–0.90 μ m. Figure 3 illustrates the acquired RGB image of the Czech data set.
(2)
Finland data: The second data set was captured over an outcrop of the Archean Siilinjärvi glimmerite-carbonatite complex in Finland, which is currently mined for large phosphate-rich apatite occurrences used in fertilizer production [55]. In the Finland data set, the same instruments were employed to acquire the HSI and RGB data. The HSI and downsampled RGB images are composed of 718 × 1848 pixels. Figure 4 displays the acquired RGB image of the Finland data set.
(3)
Trento data: The third data set was captured over a rural area in the south of the city of Trento, Italy. It consists of LiDAR and HSI data that are composed of 600 by 166 pixels with a spatial sampling of 1 m. The HSI was acquired by the AISA Eagle sensor, and contains 63 spectral bands ranging between 0.40 and 0.98 μ m. The LiDAR data were captured by the Optech ALTM 3100EA sensor. The color-composite image of the HSI from the Trento data is shown in Figure 5.
For the fusion procedure, the co-registered high-spatial resolution RGB/LiDAR images have already been downsampled to their corresponding HSIs. In the case of geological data sets, the resampling was carried out by using the nearest-neighborhood method in Quantum GIS (version 3.4, QGIS development team, Open Source Geospatial Foundation). The internal image co-registration of hyperspectral scans was done by registering all bands on a distinct master band per hypercube, using the MEPHySTo image processing toolbox [56], used for pre-processing of the hypercubes. Georeferencing of all images was done on high-resolution drone-based RGB orthophotos, created by structure-from-motion photogrammetry in Agisoft Photoscan (version 1.4, Agisoft LLC, St. Petersburg, Russia). The UAV-based, GPS-tagged Orthophotos were acquired over the target area and their positioning was refined by using ground control points. Further data acquisition details for the Czech data set are referenced in [57], and for the Finland data set in [58].

3.2. Experimental Setup

Quantitative and qualitative analyses were conducted to evaluate the performance of our proposed algorithm. The clustering results obtained by Multi-SSC with MPs and IAPs were compared with the results of two conventional clustering algorithms K-means and FCM, along with the scalable clustering algorithms, LSC [33], ESC [14], and HESSC [28]. Additionally, to allow a fair comparison, the concatenation of the HSI and RGB/LiDAR images (e.g., the combination of HSI and RGB has D + 3 features) is fed into all of the applied clustering algorithms to evaluate their performances. Moreover, to assess the effect of extracted IAPs on the clustering performance, IAPs are stacked with the spectral bands of the HSI and used as the input for the competitive clustering algorithms. For the remaining, the following acronyms are used: A L s e n s o r , where A L is the name of the applied clustering algorithm and s e n s o r refers to the applied data (HSI is only spectral data, HSI+RGB and HSI+LiDAR representing the concatenation of spectral bands and RGB/LiDAR bands, HSI+IAPs denotes the concatenation of spectral bands and IAPs).
All utilized parameters in ESC were set as the default parameters proposed by the developers [14]. In LSC, the K-means sampling strategy was used to produce the landmark points. In HESSC, the threshold τ in the lasso-based binary was set to 0.5 , the default by the developers [28]. For the MPs, suitable radii r for the SEs were selected for the first and second data sets by visual inspection. In the first data set, r = [ 1 , 5 , 10 , 15 , 20 , 40 , 60 , 80 , 100 ] were chosen. In the second data set, r = [ 10 , 20 , 40 , 60 , 80 , 100 ] , and for the last data set, r = [ 10 , 20 , 40 , 60 ] were used, as in [19]. Regarding the parameters used in extracting the IAPs, we followed the same setup as suggested by the developers in [43], where the number of scaled convolutions and their radii were set to 3 and [2,4,6], respectively. The Fourier orders to derive the frequency features were set to m = [ 0 , 1 , 2 ] , and for all data sets, we used the first three principal components to generate the IAPs.
All quantitative results are reported as an average of 10 runs. All the applied algorithms in this study were implemented in MATLAB version 2019b on a computer with an Intel© core-i9 7900X with 128 GB RAM.

3.3. Evaluation Metrics

Three commonly used evaluation metrics are employed: overall accuracy (OA), average accuracy (AA), and kappa coefficient ( κ ). The ground truth data set is represented by G = [ g 1 , g 2 , . . . , g N ] . The clustering result can be presented as M = [ m 1 , m 2 , . . . , m N ] , where m i = 1 , . . . , L , the index of the class label, assigned to each pixel. To assess the clustering performance, m i needs to be matched with g i . We thus employed m i = b e s t M a p ( g i , m i ) as a matching function, based on the Hungarian algorithm, where m i is the clustering map for which the best match between g i and m i is produced by b e s t M a p ( . ) [59]. Consequently, OA is defined as Σ i = 1 N Γ ( m i , g i ) / N , where Γ ( m i , g i ) is 1 if g i = m i and 0 otherwise.
In addition to the three aforementioned metrics, two generally applied clustering indices are employed:
(1) The adjusted rand index (ARI) is used as a similarity (or dissimilarity) measure between two partitions and it is an adjusted version of the original rand index [60]. According to [60], ARI can be formulated as:
i j n i j 2 i n i + 2 j n + j 2 / n 2 1 2 [ i n i + 2 + j n + j 2 ] i n i + 2 j n + j 2 / n 2
where, n i j = | m i g j | , n i + and n + j are equal to j = 1 N n i j and i = 1 N n i j , respectively. The range of ARI is usually between 0 and 1. An ARI closer to one means that M and G highly agree. However, it is also possible that the ARI becomes negative, indicating that the agreement between M and G is even less than what is expected from a random result.
(2) The normalized mutual information (NMI) is calculated based on the mutual information between M and G . The mutual information is normalized to range NMI between 0 and 1 [61]. NMI can be calculated as:
i j n i j log n i n i j n i + n + j ( i n i + log n i + n ) ( j n + j log n + j n )
Similar to ARI, the agreement between M and G is higher when NMI is closer to 1.

3.4. Results and Discussion

Results of all experiments are summarized in Table 1, Table 2 and Table 3, for the Czech, Finland and Trento data, respectively.

3.4.1. The Czech Data Set

The Czech data set contains 8 classes: (1) Vegetation, (2) Acid mine drainage-stream, (3) Acid mine drainage-solid, (4) Soil-slipped, (5) Clay-high, (6) Clay-flood zone, (7) Old-stream bed and (8) Artificial objects.
Table 1 summarizes all clustering results. The proposed method outperforms all other clustering approaches. The clustering performance of HESSC H S I is weak, when compared to K-means H S I , FCM H S I , and LSC H S I , which may be due to the high similarity between the spectral signatures of the different classes. The concatenation of extracted IAPs improves the accuracy of HESSC H S I + I A P s with 5% over HESSC H S I . Another 12% improvement is observed when fusing the spectral and IAPs in the proposed approach Multi-SSC, when compared to concatenating them in HESSC H S I + I A P s . With respect to the clustering results of individual classes, one can observe that class 2 and 8 are accurately mapped using Multi-SSC (IAPs), while class 4 and 6 are better mapped using Multi-SSC (MPs). ESC H S I + R G B is able to capture class 2 perfectly, while both versions of LSC perform very poor on class 2.
Figure 6 shows the obtained clustering maps (results with concatenated IAPs are not shown). It can be observed that K-means, FCM, and LSC capture the general structure, yet, both their single and multi-sensor versions generate heterogeneous clustering maps. The clustering maps obtained by Multi-SSC with MPs and IAPs are more homogeneous.
In general, the clustering results on the Czech data show that incorporating spatial information by stacking derived IAPs with the spectral bands of the HSI improves the single-sensor version of the algorithms. However, fusion of the IAPs with the HSI by employing the proposed Multi-SSC leads to the most accurate result. These results can be attributed to the structure of Multi-SSC. In the proposed fusion strategy of Multi-SSC, both spectral and spatial features contribute equally. In the competitive algorithms, i.e., K-means, FCM, LSC, and ESC, it is not possible to balance the contribution of spectral and spatial features. In the Czech data, the obtained values of ARI and NMI show that in K-means, FCM, and LSC, the mere concatenation of the IAPs, extracted from the HSI does not improve results. In ESC, however, stacking the extracted IAPs with the HSI improves ARI and NMI by almost 0.31 and 0.21, respectively. This can be attributed to the ability of ESC to select exemplars that are representative of the entire data set.

3.4.2. The Finland Data Set

The Finland data set contains eight classes: (1) Clay, (2) Glimmerite-Carbonatite, (3) Glimmerite, (4) Dark-rocks (which is a mixture of soil and Glimmerite), (5) Dust, (6) Feldspar-vein, (7) Fenite and (8) Water.
With respect to Table 2, Multi-SSC shows the highest accuracy, both with MPs and IAPs. HESSC H S I performs better than the other considered spectral-based clustering algorithms. However, the proposed algorithm (using IAPs) surpasses the original HESSC H S I by almost 17% in terms of OA. The inclusion of a high spatial-resolution RGB image in the K-means algorithm improves the clustering accuracy by 10%. Replacing RGB bands by IAPs does not further improve the results. With respect to clustering results for individual classes, ESC H S I + I A P s outperforms the other approaches on class 2 (Glimmerite-Carbonatite). However, overall, ESC H S I + I A P s has a poor performance compared to its competitors, which can be attributed to its dependence on initial conditions, i.e., the random selection of the first exemplars. HESSC H S I + I A P s and ESC H S I capture class 3 (Glimmerite) better than the other applied algorithms.
Figure 7 presents the obtained clustering maps. Multi-SSC, both with MPs and IAPs generate homogeneous regions and preserve the structures within the data set. In particular, for both geological data sets, it is desirable to obtain homogeneous clusters corresponding to individual geological features of interest. As it can be observed from Table 2, the concatenation of the extracted IAPs with the HSI in K-means, FCM, LSC, and ESC considerably increases the calculated ARI and NMI. This proves that the extracted IAPs are informative and increase the performance of the clustering algorithms. Nevertheless, the fusion strategy of Multi-SSC plays an eminent role. The equal contribution of the extracted IAPs and the HSI results in the highest ARI and NMI.

3.4.3. The Trento Data Set

In the Trento data set, there are six classes: (1) Apple trees, (2) Buildings, (3) Ground, (4) Wood, (5) Vineyard and (6) Roads. As it is shown in Table 3, also in this dataset, Multi-SSC with IAPs performs the best (OA = 71.90%) and Multi-SSC with MPs (OA = 65.12%) is the second best clustering algorithm. The clustering maps are presented in Figure 8. Overall, Multi-SSC (IAPs) generates a smooth clustering map compared to the other considered clustering algorithms. The clustering results in Table 3 show that the proposed method with IAPs miss-clustered class 1 (Apple trees) as class 5 (Vineyard). While Multi-SSC with MPs was able to cluster class 1 correctly. Such miss-clustering in Multi-SSC with IAPs can be attributed to the used parameters in the spatial extraction phase. More specifically, the used radii in the scaled convolutions lead to over-smoothed clustering results, as is shown in Figure 8m. This can be alleviated by modifying the radii of the scaled convolutions. Such modifications may, however, result in a decrease of the overall accuracy.

3.4.4. Discussion

Experimental results on all three data sets confirmed that spectral-based clustering algorithms using single-sensor data sets obtained less distinct clustering results for different classes of interest (e.g., Wood, Vineyard). The obtained values of OA, AA, κ , ARI, NMI, and t (seconds) of the applied clustering algorithms show that the inclusion of multi-sensor data sets in the clustering procedure leads to an improvement of the final clustering results. Moreover, the incorporation of spatial information using extracted spatial features IAPs versus MPs was investigated. It can be observed that the use of IAPs significantly improves the clustering performance. The results prove that the extracted IAPs consist of more informative spatial features compared to MPs, mainly due to the fact that they cover spatial information extracted from both spatial and frequency domains.
Multi-SSC was designed to fuse the rich spectral, spatial, and contextual information derived from multi-sensor data sets. The algorithm has the potential capability to be adjusted for processing any type of data set. Multi-SSC benefits from a hierarchical structure, which makes the algorithm able to describe multi-sensor data sets at different levels of detail. In addition, Multi-SSC employs the sparse representation matrix to cluster data sets. Such information allows the algorithm to explore and identify the relation between data points from the same subspace. Due to its effective structure, Multi-SSC is able to handle large scale data sets by using limited number of data points. Multi-SSC utilizes both spectral and spatial information in order to generate the final clustering map. As can be seen in Table 1, Table 2 and Table 3, the best results are obtained with the proposed algorithm. Multi-SSC is, thus, able to fuse and exploit spatial information in more accurate manner compared to simply concatenating two data sets. In addition, Multi-SSC ensures a balanced use of the spectral and spatial information in its clustering procedure.
Therefore, it can be summarized that Multi-SSC, both using MPs and IAPs, performs well to capture the classes of interest in both rural and geological areas. Multi-SSC consistently performs better with IAPs than with MPs. This can be attributed to the fixed shape of the SEs in MPs, which cannot fully exploit the spatial characteristics of regions within a scene of an image. Additionally, due to the fixed parameters, IAPs can be automatically extracted, which makes them more efficient compared to MPs for unsupervised learning purposes.
Processing times are reported in Table 1, Table 2 and Table 3. Overall, K-means and LSC perform fast, compared to the other applied clustering algorithms. K-means is a distance-based algorithm that measures the Euclidean distance between each data point and its corresponding centroid, which does not demand high computational power. In LSC, landmark points are identified by adopting the K-means algorithm, leading to similar processing times. Also, based on the reported ARIs and NMIs on different data sets, K-means and LSC performed similarly. According to the reported processing time for all data sets, Multi-SSC, both with MPs and IAPs takes more time than the other approaches K-means, FCM, LSC, and HESSC. This can be attributed to (1) the spatial feature extraction step, which can be time consuming when extracting the spatial information from larger data sets, and (2) the ECC step, which generates the final binary clustering result from a binary matrix that is composed of 100 individual binary clustering results) [28,51]. Nonetheless, with respect to the calculated ARI and NMI, the most accurate performance on all data sets is obtained by Multi-SSC with IAPs. This observation shows that our proposed algorithm is able to effectively fuse the spatial/contextual and spectral information from multiple-sensor data sets. As another observation, among all the applied clustering algorithms, ESC performs the slowest. This can be attributed to the high number of exemplars selected by ESC. By increasing the number of exemplars, not only the processing step will increase, but also the capability of ESC to handle large-scale data sets will drastically drop.

4. Conclusions

In this paper, we proposed Multi-SSC, a novel multi-sensor clustering algorithm that integrates high spectral and spatial resolution images. In Multi-SSC, invariant attribute profiles (IAPs) were extracted to incorporate the spatial information derived from a fine spatial-resolution image (RGB/ LiDAR). Then, the hierarchical sparse subspace clustering algorithm (HESSC) was adopted to fuse the spectral information derived from an HSI with the extracted IAPs to produce the final clustering map. The clustering performance of Multi-SSC was validated on three different data sets. Two data sets were acquired over mining sites and used to identify the minerals of interest. The third data set is the well-known Trento benchmark data set containing HSI and LiDAR data over a rural area. Multi-SSC along with some conventional and state-of-the-art subspace clustering algorithms were applied on all three data sets.
The experiments imply that in most cases, the concatenation of IAPs with spectral bands of an HSI has the superiority to merely stacking the HSI bands with RGB and LiDAR bands. Furthermore, it is also observed that, fusing the multi sensor data sets using Multi-SSC considerably increases the clustering accuracies. The clustering maps obtained by Multi-SSC contain homogeneous areas and preserve the structures present in the scene. The proposed multi-sensor algorithm evenly balances the contribution of spectral and spatial information, and allows to adapt the ratio of employing spectral to spatial information according to the applications at hand.

Author Contributions

Conceptualization, K.R.S., P.G. and B.R.; methodology, K.R.S. and B.R.; software, K.R.S.; validation, K.R.S. and R.J.; formal analysis, K.R.S.; investigation, K.R.S.; writing—original draft preparation, K.R.S.; writing—review and editing, all authors; visualization, K.R.S.; supervision, P.G., P.S. and R.G.; project administration, R.G. All authors have read and agreed to the published version of the manuscript.

Funding

This study received funding from the Federal Ministry of Education and Research (BMBF), client II program, within the MoCa project under grant assignment No.033R189B. More information regarding the project can be found in the following link (https://www.bmbf-client.de/en/projects/moca). The work of Behnood Rasti was funded by Alexander von Humboldt foundation.

Acknowledgments

The authors would like to thank Lorezone Bruzzone of the University of Trento for providing the Trento data set. The authors would like to acknowledge the Federal Ministry of Education and Research (BMBF) for funding this study, in the client II program, within the MoCa project under grant assignment No.033R189B.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Meng, T.; Jing, X.; Yan, Z.; Pedrycz, W. A survey on machine learning for data fusion. Inf. Fusion 2020, 57, 115–129. [Google Scholar] [CrossRef]
  2. Goetz, A.F.; Vane, G.; Solomon, J.E.; Rock, B.N. Imaging Spectrometry for Earth Remote Sensing. Science 1985, 228, 1147–1153. [Google Scholar] [CrossRef] [PubMed]
  3. Orynbaikyzy, A.; Gessner, U.; Conrad, C. Crop type classification using a combination of optical and radar remote sensing data: A review. Int. J. Remote Sens. 2019, 40, 6553–6595. [Google Scholar] [CrossRef]
  4. Ghamisi, P.; Rasti, B.; Yokoya, N.; Wang, Q.; Hofle, B.; Bruzzone, L.; Bovolo, F.; Chi, M.; Anders, K.; Gloaguen, R.; et al. Multisource and Multitemporal Data Fusion in Remote Sensing: A Comprehensive Review of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2019, 7, 6–39. [Google Scholar] [CrossRef] [Green Version]
  5. Lorenz, S.; Seidel, P.; Ghamisi, P.; Zimmermann, R.; Tusa, L.; Khodadadzadeh, M.; Contreras, I.C.; Gloaguen, R. Multi-Sensor Spectral Imaging of Geological Samples: A Data Fusion Approach Using Spatio-Spectral Feature Extraction. Sensors 2019, 19, 2787. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Tusa, L.; Andreani, L.; Khodadadzadeh, M.; Contreras, C.; Ivascanu, P.; Gloaguen, R.; Gutzmer, J. Mineral Mapping and Vein Detection in Hyperspectral Drill-Core Scans: Application to Porphyry-Type Mineralization. Minerals 2019, 9, 122. [Google Scholar] [CrossRef] [Green Version]
  7. Cho, M.A.; Skidmore, A.; Corsi, F.; van Wieren, S.E.; Sobhan, I. Estimation of green grass/herb biomass from airborne hyperspectral imagery using spectral indices and partial least squares regression. Int. J. Appl. Earth Obs. Geoinf. 2007, 9, 414–424. [Google Scholar] [CrossRef]
  8. Tuşa, L.; Khodadadzadeh, M.; Contreras, C.; Rafiezadeh Shahi, K.; Fuchs, M.; Gloaguen, R.; Gutzmer, J. Drill-Core Mineral Abundance Estimation Using Hyperspectral and High-Resolution Mineralogical Data. Remote Sens. 2020, 12, 1218. [Google Scholar] [CrossRef] [Green Version]
  9. Hughes, G. On the mean accuracy of statistical pattern recognizers. IEEE Trans. Inf. Theory 1968, 55–63. [Google Scholar] [CrossRef] [Green Version]
  10. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in Hyperspectral Image and Signal Processing: A Comprehensive Overview of the State of the Art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef] [Green Version]
  11. Ghamisi, P.; Maggiori, E.; Li, S.; Souza, R.; Tarablaka, Y.; Moser, G.; De Giorgi, A.; Fang, L.; Chen, Y.; Chi, M.; et al. New Frontiers in Spectral-Spatial Hyperspectral Image Classification: The Latest Advances Based on Mathematical Morphology, Markov Random Fields, Segmentation, Sparse Representation, and Deep Learning. IEEE Geosci. Remote Sens. Mag. 2018, 6, 10–43. [Google Scholar] [CrossRef]
  12. Ghamisi, P.; Plaza, J.; Chen, Y.; Li, J.; Plaza, A.J. Advanced Spectral Classifiers for Hyperspectral Images: A review. IEEE Geosci. Remote Sens. Mag. 2017, 5, 8–32. [Google Scholar] [CrossRef] [Green Version]
  13. He, L.; Li, J.; Liu, C.; Li, S. Recent Advances on Spectral–Spatial Hyperspectral Image Classification: An Overview and New Guidelines. IEEE Trans. Geosci. Remote Sens. 2018, 56, 1579–1597. [Google Scholar] [CrossRef]
  14. You, C.; Li, C.; Robinson, D.P.; Vidal, R. Self-Representation Based Unsupervised Exemplar Selection in a Union of Subspaces. arXiv 2020, arXiv:2006.04246. [Google Scholar] [CrossRef] [PubMed]
  15. Arthur, D.; Vassilvitskii, S. k-Means plus plus: The advantages of careful seeding. In Proceedings of the Eighteenth Annual Acm-Siam Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; pp. 1027–1035. [Google Scholar]
  16. Bezdek, J.C. Pattern Recognition with Fuzzy Objective Function Algorithms; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  17. Elhamifar, E.; Vidal, R. Sparse Subspace Clustering: Algorithm, Theory, and Applications. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2765–2781. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Li, H.; Ghamisi, P.; Soergel, U.; Zhu, X.X. Hyperspectral and lidar fusion using deep three-stream convolutional neural networks. Remote Sens. 2018, 10, 1649. [Google Scholar] [CrossRef] [Green Version]
  19. Rasti, B.; Ghamisi, P. Remote sensing image classification using subspace sensor fusion. Inf. Fusion 2020, 64, 121–130. [Google Scholar] [CrossRef]
  20. Rasti, B.; Ghamisi, P.; Gloaguen, R. Hyperspectral and LiDAR Fusion Using Extinction Profiles and Total Variation Component Analysis. IEEE Trans. Geosci. Remote Sens. 2017, 55, 3997–4007. [Google Scholar] [CrossRef]
  21. Hartling, S.; Sagan, V.; Sidike, P.; Maimaitijiang, M.; Carron, J. Urban Tree Species Classification Using a WorldView-2/3 and LiDAR Data Fusion Approach and Deep Learning. Sensors 2019, 19, 1284. [Google Scholar] [CrossRef] [Green Version]
  22. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and Multispectral Data Fusion: A comparative review of the recent literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  23. Pelizari, P.A.; Spröhnle, K.; Geiß, C.; Schoepfer, E.; Plank, S.; Taubenböck, H. Multi-sensor feature fusion for very high spatial resolution built-up area extraction in temporary settlements. Remote Sens. Environ. 2018, 209, 793–807. [Google Scholar] [CrossRef]
  24. Sellami, A.; Abbes, A.B.; Barra, V.; Farah, I.R. Fused 3-D spectral-spatial deep neural networks and spectral clustering for hyperspectral image classification. Pattern Recognit. Lett. 2020, 138, 594–600. [Google Scholar] [CrossRef]
  25. Pourkamali-Anaraki, F.; Folberth, J.; Becker, S. Efficient Solvers for Sparse Subspace Clustering. Signal Process. 2020, 172, 107548. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, H.; Zhai, H.; Zhang, L.; Li, P. Spectral–Spatial Sparse Subspace Clustering for Hyperspectral Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3672–3684. [Google Scholar] [CrossRef]
  27. You, C.; Vidal, R. Sparse Subspace Clustering by Orthogonal Matching Pursuit. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3918–3927. [Google Scholar]
  28. Rafiezadeh Shahi, K.; Khodadadzadeh, M.; Tusa, L.; Ghamisi, P.; Tolosana-Delgado, R.; Gloaguen, R. Hierarchical Sparse Subspace Clustering (HESSC): An Automatic Approach for Hyperspectral Image Analysis. Remote Sens. 2020, 12, 2421. [Google Scholar] [CrossRef]
  29. Zhang, Q.; Liu, Y.; Blum, R.S.; Han, J.; Tao, D. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Inf. Fusion 2018, 40, 57–75. [Google Scholar] [CrossRef]
  30. Shahi, K.R.; Khodadadzadeh, M.; Tolosana-delgado, R.; Tusa, L.; Gloaguen, R. The Application Of Subspace Clustering Algorithms In Drill-Core Hyperspectral Domaining. In Proceedings of the 2019 10th Workshop on Hyperspectral Imaging and Signal Processing: Evolution in Remote Sensing (WHISPERS), Amsterdam, The Netherlands, 24–26 September 2019; pp. 1–5. [Google Scholar] [CrossRef]
  31. Rafiezadeh Shahi, K.; Ghamisi, P.; Jackisch, R.; Khodadadzadeh, M.; Lorenz, S.; Gloaguen, R. A new spectral-spatial subspace clustering algorithm for hyperspectral image analysis. ISPRS Ann. Photogramm. Remote. Sens. Spat. Inf. Sci. 2020, 3, 185–191. [Google Scholar] [CrossRef]
  32. Vidal, R. Subspace Clustering. IEEE Signal Process. Mag. 2011, 28, 52–68. [Google Scholar] [CrossRef]
  33. Cai, D.; Chen, X. Large scale spectral clustering via landmark-based sparse representation. IEEE Trans. Cybern. 2014, 45, 1669–1680. [Google Scholar]
  34. Dyer, E.; Sankaranarayanan, A.; Baraniuk, R. Greedy feature selection for subspace clustering. J. Mach. Learn. Res. 2013, 14, 2487–2517. [Google Scholar]
  35. Guo, Y.; Gao, J.; Li, F. Spatial subspace clustering for drill hole spectral data. J. Appl. Remote Sens. 2014, 8, 1–20. [Google Scholar] [CrossRef]
  36. Hinojosa, C.; Bacca, J.; Arguello, H. Coded Aperture Design for Compressive Spectral Subspace Clustering. IEEE J. Sel. Top. Signal Process. 2018, 12, 1589–1600. [Google Scholar] [CrossRef]
  37. Pesaresi, M.; Benediktsson, J.A. A new approach for the morphological segmentation of high-resolution satellite imagery. IEEE Trans. Geosci. Remote Sens. 2001, 39, 309–320. [Google Scholar] [CrossRef] [Green Version]
  38. Benediktsson, J.A.; Palmason, J.A.; Sveinsson, J.R. Classification of hyperspectral data from urban areas based on extended morphological profiles. IEEE Trans. Geosci. Remote Sens. 2005, 43, 480–491. [Google Scholar] [CrossRef]
  39. Dalla Mura, M.; Benediktsson, J.A.; Waske, B.; Bruzzone, L. Morphological Attribute Profiles for the Analysis of Very High Resolution Images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3747–3762. [Google Scholar] [CrossRef]
  40. Dalla Mura, M.; Villa, A.; Benediktsson, J.A.; Chanussot, J.; Bruzzone, L. Classification of Hyperspectral Images by Using Extended Morphological Attribute Profiles and Independent Component Analysis. IEEE Geosci. Remote Sens. Lett. 2011, 8, 542–546. [Google Scholar] [CrossRef] [Green Version]
  41. Huang, X.; Guan, X.; Benediktsson, J.A.; Zhang, L.; Li, J.; Plaza, A.; Dalla Mura, M. Multiple Morphological Profiles From Multicomponent-Base Images for Hyperspectral Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 4653–4669. [Google Scholar] [CrossRef]
  42. Ghamisi, P.; Dalla Mura, M.; Benediktsson, J.A. A Survey on Spectral–Spatial Classification Techniques Based on Attribute Profiles. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2335–2353. [Google Scholar] [CrossRef]
  43. Hong, D.; Wu, X.; Ghamisi, P.; Chanussot, J.; Yokoya, N.; Zhu, X.X. Invariant Attribute Profiles: A Spatial-Frequency Joint Feature Extractor for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3791–3808. [Google Scholar] [CrossRef] [Green Version]
  44. Liu, K.; Skibbe, H.; Schmidt, T.; Blein, T.; Palme, K.; Brox, T.; Ronneberger, O. Rotation-invariant HOG descriptors using Fourier analysis in polar and spherical coordinates. Int. J. Comput. Vis. 2014, 106, 342–364. [Google Scholar] [CrossRef]
  45. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC Superpixels Compared to State-of-the-Art Superpixel Methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [Green Version]
  46. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar]
  47. Von Luxburg, U. A tutorial on spectral clustering. Stat. Comput. 2007, 17, 395–416. [Google Scholar] [CrossRef]
  48. Ng, A.Y.; Jordan, M.I.; Weiss, Y. On spectral clustering: Analysis and an algorithm. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 9–14 December 2002; pp. 849–856. [Google Scholar]
  49. Chen, X.; Cai, D. Large scale spectral clustering with landmark-based representation. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011. [Google Scholar]
  50. Efron, B.; Hastie, T.; Johnstone, I.; Tibshirani, R. Least angle regression. Ann. Statist. 2004, 32, 407–499. [Google Scholar]
  51. Liu, H.; Zhao, R.; Fang, H.; Cheng, F.; Fu, Y.; Liu, Y.Y. Entropy-based consensus clustering for patient stratification. Bioinformatics 2017, 33, 2691–2698. [Google Scholar] [CrossRef]
  52. Wu, T.; Gurram, P.; Rao, R.M.; Bajwa, W.U. Hierarchical union-of-subspaces model for human activity summarization. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 11–12 December 2015; pp. 1–9. [Google Scholar]
  53. Gu, Y.; Zhang, Y.; Zhang, J. Integration of Spatial–Spectral Information for Resolution Enhancement in Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1347–1358. [Google Scholar]
  54. Gillmann, C.; Arbelaez, P.; Hernandez, J.T.; Hagen, H.; Wischgoll, T. An uncertainty-aware visual system for image pre-processing. J. Imaging 2018, 4, 109. [Google Scholar] [CrossRef] [Green Version]
  55. O’Brien, H.; Heilimo, E.; Heino, P. Chapter 4.3—The Archean Siilinjärvi Carbonatite Complex. In Mineral Deposits of Finland; Maier, W.D., Lahtinen, R., O’Brien, H., Eds.; Elsevier: Amsterdam, The Netherlands, 2015; pp. 327–343. [Google Scholar] [CrossRef]
  56. Jakob, S.; Zimmermann, R.; Gloaguen, R. The Need for Accurate Geometric and Radiometric Corrections of Drone-Borne Hyperspectral Data for Mineral Exploration: MEPHySTo—A Toolbox for Pre-Processing Drone-Borne Hyperspectral Data. Remote Sens. 2017, 9, 88. [Google Scholar] [CrossRef] [Green Version]
  57. Jackisch, R.; Lorenz, S.; Zimmermann, R.; Möckel, R.; Gloaguen, R. Drone-Borne Hyperspectral Monitoring of Acid Mine Drainage: An Example from the Sokolov Lignite District. Remote Sens. 2018, 10, 385. [Google Scholar] [CrossRef] [Green Version]
  58. Jackisch, R.; Lorenz, S.; Kirsch, M.; Zimmermann, R.; Tusa, L.; Pirttijaervi, M.; Saartenoja, A.; Ugalde, H.; Madriz, Y.; Savolainen, M.; et al. Integrated Geological and Geophysical Mapping of a Carbonatite-Hosting Outcrop in Siilinjärvi, Finland, Using Unmanned Aerial Systems. Remote Sens. 2020, 12, 2998. [Google Scholar] [CrossRef]
  59. Rezaei, M.; Fränti, P. Set Matching Measures for External Cluster Validity. IEEE Trans. Knowl. Data Eng. 2016, 28, 2173–2186. [Google Scholar] [CrossRef]
  60. Hubert, L.; Arabie, P. Comparing partitions. J. Classif. 1985, 2, 193–218. [Google Scholar] [CrossRef]
  61. Wu, J.; Xiong, H.; Chen, J. Adapting the Right Measures for K-means Clustering. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; pp. 877–886. [Google Scholar] [CrossRef]
Figure 1. The proposed multi-sensor sparse-based clustering algorithm for HSI analysis.
Figure 1. The proposed multi-sensor sparse-based clustering algorithm for HSI analysis.
Remotesensing 12 04007 g001
Figure 2. The lasso-based binary algorithm employed in hierarchical sparse subspace clustering algorithm (HESSC) in one iteration.
Figure 2. The lasso-based binary algorithm employed in hierarchical sparse subspace clustering algorithm (HESSC) in one iteration.
Remotesensing 12 04007 g002
Figure 3. RGB image from the Czech data in which the HSI and RGB images were captured over a tailing area with an adjacent lake near Litov in Sokolov, Czech Republic.
Figure 3. RGB image from the Czech data in which the HSI and RGB images were captured over a tailing area with an adjacent lake near Litov in Sokolov, Czech Republic.
Remotesensing 12 04007 g003
Figure 4. RGB image from the Finland data in which the HSI and RGB images were captured at Siilinjärvi in Finland.
Figure 4. RGB image from the Finland data in which the HSI and RGB images were captured at Siilinjärvi in Finland.
Remotesensing 12 04007 g004
Figure 5. False color-composite image of the HSI from the Trento data using bands R:40, G:20, B:10.
Figure 5. False color-composite image of the HSI from the Trento data using bands R:40, G:20, B:10.
Remotesensing 12 04007 g005
Figure 6. The Czech data. (a) RGB image, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + R G B , (e) FCM H S I , (f) FCM H S I + R G B , (g) LSC H S I , (h) LSC H S I + R G B , (i) ESC H S I , (j) ESC H S I + R G B , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Figure 6. The Czech data. (a) RGB image, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + R G B , (e) FCM H S I , (f) FCM H S I + R G B , (g) LSC H S I , (h) LSC H S I + R G B , (i) ESC H S I , (j) ESC H S I + R G B , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Remotesensing 12 04007 g006
Figure 7. The Finland data. (a) RGB image, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + R G B , (e) FCM H S I , (f) FCM H S I + R G B , (g) LSC H S I , (h) LSC H S I + R G B , (i) ESC H S I , (j) ESC H S I + R G B , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Figure 7. The Finland data. (a) RGB image, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + R G B , (e) FCM H S I , (f) FCM H S I + R G B , (g) LSC H S I , (h) LSC H S I + R G B , (i) ESC H S I , (j) ESC H S I + R G B , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Remotesensing 12 04007 g007
Figure 8. The Trento data. (a) HSI, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + L i D A R , (e) FCM H S I , (f) FCM H S I + L i D A R , (g) LSC H S I , (h) LSC H S I + L i D A R , (i) ESC H S I , (j) ESC H S I + L i D A R , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Figure 8. The Trento data. (a) HSI, (b) ground truth, and clustering maps obtained by (c) K-means H S I , (d) K-means H S I + L i D A R , (e) FCM H S I , (f) FCM H S I + L i D A R , (g) LSC H S I , (h) LSC H S I + L i D A R , (i) ESC H S I , (j) ESC H S I + L i D A R , (k) HESSC H S I , (l) Multi-SSC (MPs), (m) Multi-SSC (IAPs).
Remotesensing 12 04007 g008
Table 1. Quantitative assessment of the performances of the clustering algorithms applied to the Czech data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
Table 1. Quantitative assessment of the performances of the clustering algorithms applied to the Czech data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
ClassNo. Ground Truth SamplesK-means HSI K-means HSI + RGB K-means HSI + IAPs FCM HSI FCM HSI + RGB FCM HSI + IAPs LSC HSI LSC HSI + RGB LSC HSI + IAPs ESC HSI ESC HSI + RGB ESC HSI + IAPs HESSC HSI HESSC HSI + IAPs Multi-SSC (MPs)Multi-SSC (IAPs)
1101850.4987.9291.0636.0584.8744.4010093.4280.3510099.7198.0463.0689.0055.5097.05
290198.6793.0188.6797.3410.5466.890.000.1199.0099.0099.6753.5066.2626.5377.3699.67
3111973.7371.4071.9463.8171.6746.8550.6779.5467.7441.820.5441.8222.880.0051.0336.73
487452.860.003.4365.9029.1860.6464.3016.8265.220.572.1794.1647.9486.6180.3256.86
583897.6199.0587.6196.0698.6986.4295.9497.020.1241.7780.190.0065.7568.5056.4485.68
686331.2966.7468.2558.8642.5359.1035.5742.4147.740.1286.330.0012.0598.7386.7929.66
777738.7449.5535.3935.9146.7235.6535.3955.0833.3310.290.0011.4565.8916.2228.0632.95
87850.007.3931.974.592.554.598.5419.759.9443.9551.7279.3643.8238.8517.7180.13
OA56.8561.0661.3758.0050.0856.7050.1752.2855.6843.0552.4245.4547.7452.1357.3464.85
AA55.4259.3859.7957.3148.3450.5748.8050.5150.4342.1952.5447.2948.4553.0556.6564.84
κ 0.500.550.580.510.420.480.420.450.450.340.450.480.400.460.510.59
A R I 0.500.470.490.410.420.360.370.450.330.220.360.530.310.350.530.54
N M I 0.630.620.630.540.560.510.530.580.510.410.550.620.500.520.640.66
t (seconds)1.161.011.66145.12187.09196.6234.6534.9435.0113,387.009859.0011,130.003562.803557.203086.513114.80
Table 2. Quantitative assessment of the performances of the clustering algorithms applied to the Finland data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
Table 2. Quantitative assessment of the performances of the clustering algorithms applied to the Finland data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
ClassNo. Ground Truth SamplesK-means HSI K-means HSI + RGB K-means HSI + IAPs FCM HSI FCM HSI + RGB FCM HSI + IAPs LSC HSI LSC HSI + RGB LSC HSI + IAPs ESC HSI ESC HSI + RGB ESC HSI + IAPs HESSC HSI HESSC HSI + IAPs Multi-SSC (MPs)Multi-SSC (IAPs)
1106253.5854.4371.2848.9649.8166.2068.1750.8531.1777.7895.6713.7554.9075.3356.97100
27910.0025.9220.860.000.000.003.921.649.991.391.9041.0922.124.238.342.28
3104854.4865.5561.4531.5832.4455.1541.7037.4052.8671.280.007.630.0089.5053.3471.18
499450.8094.1655.3343.8644.4752.1147.3856.9466.900.000.800.0096.3824.1485.9285.01
596456.6445.3365.6661.5162.1468.6740.8739.6343.571.040.0045.750.217.573.0112.34
6106188.2269.1887.7586.9086.7188.030.0089.4494.250.1930.7326.8642.7982.7583.6971.25
7106535.8758.590.0037.8438.1211.8770.0530.4216.4234.4637.4694.9657.4615.8398.2256.81
810110.0010.2968.8337.2938.6757.2885.6629.2837.9310010019.1588.8233.7149.1191.10
OA43.8853.8453.1544.8045.3651.6245.8943.3044.6837.1934.7040.7146.0544.5656.4963.43
AA42.4552.9353.9043.4944.0549.9144.7241.9544.1435.7733.3231.1545.3441.6354.8261.25
κ 0.350.470.480.360.370.440.380.350.370.270.240.210.380.420.500.58
A R I 0.270.380.320.280.270.360.300.270.330.160.110.280.340.230.410.46
N M I 0.420.520.490.430.420.510.490.420.510.370.350.380.510.420.550.57
t (seconds)27.8154.5655.45318.40339.15338.1852.0557.7171.11100,890.00100,210.00110,700.0012,085.0070,492.0090,436.0098,752.00
Table 3. Quantitative assessment of the performances of the clustering algorithms applied to the Trento data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
Table 3. Quantitative assessment of the performances of the clustering algorithms applied to the Trento data. The clustering accuracy is reported using OA, AA, κ , ARI, and NMI. In addition, the processing time is reported as well.
ClustersNo. Ground Truth SamplesK-means HSI K-means HSI + LiDAR K-means HSI + IAPs FCM HSI FCM HSI + LiDAR FCM HSI + IAPs LSC HSI LSC HSI + LiDAR LSC HSI + IAPs ESC HSI ESC HSI + LiDAR ESC HSI + IAPs HESSC HSI HESSC HSI + IAPs Multi-SSC (MPs)Multi-SSC (IAPs)
1403470.9138.3459.6775.9024.8655.3560.1343.2755.3310.5526.822.780.0045.7029.300.00
2290389.2598.970.5584.4699.440.650.0035.592.1710.0899.2264.1780.2681.4636.3368.55
34790.0015.6430.9059.4659.6732.3610.7716.1624.8442.784.1363.6721.7125.659.6026.72
4912357.0786.3599.930.0069.8699.9156.7068.8299.6234.3574.1554.5891.6970.6599.9590.57
510,50142.7948.6869.0539.1822.4268.5035.8136.8860.0169.3744.0573.3243.7039.1159.2694.47
6317431.2315.7252.9653.097.6353.0294.7290.6570.0183.8146.8450.5055.4534.0082.8944.83
OA51.9759.5263.2253.8854.2158.6547.7152.9663.7646.3463.9554.1356.7652.0965.1271.90
AA48.5450.6152.1851.1844.7751.6343.0249.2352.0041.8249.2051.5048.8049.4352.8954.19
κ 0.460.460.530.430.420.540.370.450.540.270.500.510.410.490.550.61
A R I 0.270.160.500.280.160.510.280.180.420.210.300.320.370.360.440.53
N M I 0.430.250.540.460.250.540.450.260.560.370.440.380.490.460.580.64
t (seconds)2.552.733.0121.6920.4211.603.862.482.96763.52118.92764.79478.94576.11407.49518.63
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rafiezadeh Shahi, K.; Ghamisi, P.; Rasti, B.; Jackisch, R.; Scheunders, P.; Gloaguen, R. Data Fusion Using a Multi-Sensor Sparse-Based Clustering Algorithm. Remote Sens. 2020, 12, 4007. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12234007

AMA Style

Rafiezadeh Shahi K, Ghamisi P, Rasti B, Jackisch R, Scheunders P, Gloaguen R. Data Fusion Using a Multi-Sensor Sparse-Based Clustering Algorithm. Remote Sensing. 2020; 12(23):4007. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12234007

Chicago/Turabian Style

Rafiezadeh Shahi, Kasra, Pedram Ghamisi, Behnood Rasti, Robert Jackisch, Paul Scheunders, and Richard Gloaguen. 2020. "Data Fusion Using a Multi-Sensor Sparse-Based Clustering Algorithm" Remote Sensing 12, no. 23: 4007. https://0-doi-org.brum.beds.ac.uk/10.3390/rs12234007

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop