Next Article in Journal
GNSS-R Soil Moisture Retrieval for Flat Vegetated Surfaces Using a Physics-Based Bistatic Scattering Model and Hybrid Global/Local Optimization
Previous Article in Journal
Hydrological Evaluation of Satellite-Based Precipitation Products in Hunan Province
Previous Article in Special Issue
Registration of Multisensor Images through a Conditional Generative Adversarial Network and a Correlation-Type Similarity Measure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Densely Residual Network with Dual Attention for Hyperspectral Reconstruction from RGB Images

1
Xiaomi, Nanjing 210019, China
2
Department of Computer Science, Faculty of Information Technology and Electrical Engineering, Norwegian University of Science and Technology, N-2815 Gjøvik, Norway
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(13), 3128; https://0-doi-org.brum.beds.ac.uk/10.3390/rs14133128
Submission received: 5 March 2022 / Revised: 7 June 2022 / Accepted: 10 June 2022 / Published: 29 June 2022
(This article belongs to the Special Issue Machine Learning for Remote Sensing Image/Signal Processing)

Abstract

:
In the last several years, deep learning has been introduced to recover a hyperspectral image (HSI) from a single RGB image and demonstrated good performance. In particular, attention mechanisms have further strengthened discriminative features, but most of them are learned by convolutions with limited receptive fields or require much computational cost, which hinders the function of attention modules. Furthermore, the performance of these deep learning methods is hampered by tackling multi-level features equally. To this end, in this paper, based on multiple lightweight densely residual modules, we propose a densely residual network with dual attention (DRN-DA), which utilizes advanced attention and adaptive fusion strategy for more efficient feature correlation learning and more powerful feature extraction. Specifically, an SE layer is applied to learn channel-wise dependencies, and dual downsampling spatial attention (DDSA) is developed to capture long-range spatial contextual information. All the intermediate-layer feature maps are adaptively fused. Experimental results on four data sets from the NTIRE 2018 and NTIRE 2020 Spectral Reconstruction Challenges demonstrate the superiority of the proposed DRN-DA over state-of-the-art methods (at least −6.19% and −1.43% on NTIRE 2018 “Clean” track and “Real World” track, −6.85% and −5.30% on NTIRE 2020 “Clean” track and “Real World” track) in terms of mean relative absolute error.

Graphical Abstract

1. Introduction

With the increasing applications of computer vision technology in various engineering fields [1,2,3,4,5,6,7,8,9,10,11], hyperspectral images (HSIs) have proved to obtain more helpful information than RGB images. Hyperspectral images contain the reflectance of objects or scenes in different spectral bands, usually ranging from several dozens to hundreds, even outside the visible spectrum (e.g., in the ultraviolet or infrared spectrum). Compared with traditional RGB images with increased spectral range and resolution, HSIs provide much richer information, which has been widely used in cultural heritage [1,2], medical diagnosis [3], remote sensing [4], food quality inspection [5], color quality control [6], and various computer vision tasks, such as face recognition, object tracking, and image classification [7,8,9].
Due to the growing need for HSIs, various hyperspectral imaging systems (HISs) have been developed in the last several decades. The first HISs, such as NASA’s airborne visible/infrared imaging spectrometer (AVIRIS) [12], employed a prism to disperse the reflected light and a linear array detector to record the reflected light. This kind of HISs can acquire images with high spatial/spectral resolution in “whisk broom” imaging mode, but image acquisition is time-consuming since they adopt the point-scanning method. Afterwards, “push broom” HISs, such as NASA’s advanced land imager (ALI) [13], and “staring” HISs, often used in microscopy or other lab applications, have been developed. With a scan lens and an entrance slit, “push broom” HISs use a prism-grating-prism to split the light and can scan one line at a time; “staring” HISs use continuously changeable narrow bandpass filters in front of a matrix detector and can capture the image at one wavelength at a given time. Obviously, the latter two have relatively fast speed of scanning, but their temporal resolutions are still not high enough for dynamic scenes. Thus, many non-scanning, snapshot HISs have been designed, including computed tomographic imaging spectrometry (CTIS), fiber-reformatting imaging spectrometry (FRIS), and coded aperture snapshot spectral imaging (CASSI) [14]. Unfortunately, as the total number of voxels cannot exceed the total number of pixels on the CCD camera, these HISs involve a trade-off between spatial and spectral resolution [15].
As aforementioned, HISs to directly acquire HSIs have disadvantages either in temporal resolution or in spatial resolution. In recent years, hyperspectral reconstruction from RGB images has become a very active research topic. A large number of methods have been proposed to reconstruct hyperspectral information using only RGB cameras [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]. In general, these methods fall into three branches: traditional, machine learning, and deep learning methods. Traditional methods include classical methods on the basis of Wiener estimation, pseudo-inverse estimation, or principal component analysis, and their various modifications, such as adaptive Wiener estimation [16], regularized local linear models [17], sequential weighted nonlinear regression models [18], and so on. Classical methods are simple and straight but not very accurate. Modified methods tend to adaptively select or weigh training samples, which enhances hyperspectral reconstruction accuracy, but are not portable and can be time-consuming. Therefore, it is hard for these traditional methods to be applied in real tasks. Fortunately, machine learning-based methods compute hyperspectral data, which makes acquisition of spectral data fast. Shallow learning models such as RBF networks [19,20], dictionary learning-based sparse coding [21,22], and manifold-based mapping [23] are typical machine learning-based methods. Nevertheless, the expression capacity of handcrafted prior models learned by these methods is so limited that the reconstructed spectra are not very accurate. In recent years, deep learning methods have been used for hyperspectral reconstruction and have achieved remarkable success. Compared to machine learning methods, deep learning methods are capable of automatically extracting high-level features and have a better generalization ability. Convolutional neural networks (CNNs) [24,25,26,27,28,29,30,31] and generative adversarial networks (GANs) [32,33] have been developed for hyperspectral reconstruction.
Though considerable progress has been made for hyperspectral reconstruction, challenges remain in deep learning-based models as noted below: (1) when using dense skip connections, the feature of each layer is propagated into all subsequent layers, resulting in a very wide network at the cost of reducing the depth of the network [34]; (2) the commonly used non-local spatial attention involves big matrix multiplications, raising the cost of computation and memory requirement, thus hindering the frequent application in the network [35]; (3) deep learning-based models for hyperspectral reconstruction neglect that the importance of different intermediate layers varies. To address these issues, we propose a densely residual network with dual attention (DRN-DA) for more powerful feature representation, which adequately enjoys the benefits of both the residual block [36] and the dense block [34]. In our proposed DRN-DA network, the basic building blocks are densely residual block (DRB) and densely residual attention block (DRAB). The difference between DRB and DRAB is that DRAB has channel attention (CA) and dual downsampling spatial attention (DDSA) to capture channel-wise and long-range spatial dependencies. Then, to reuse the features, the output features of DRB and DRAB are adaptively fused. Additionally, an adaptive multi-scale block (AMB) with larger receptive fields is used to process the features generated by the previous network. Extensive experiments demonstrate that the proposed DRN-DA network performs better when compared to the state-of-the-art methods.
In summary, the contributions of this paper are as follows:
  • We propose a novel model, named densely residual network with dual attention (DRN-DA), which enhances the representation ability of feature learning for hyperspectral reconstruction.
  • We propose a lightweight dense skip connection, where each layer is connected to the next layer rather than all the subsequent layers. Although this block is different from the classic DenseNet [34], it also reuses features and eliminates gradient vanishing.
  • We propose a simple but effective non-local block named dual downsampling spatial attention (DDSA) to decrease the computation and memory consumption of the standard non-local block, which makes it feasible to insert multiple non-local blocks in the network for enhancing the performance.
  • To further improve the learning ability of the network, we introduce an adaptive fusion block (AFB) to adaptively reuse the features from different intermediate layers.
The remainder of this paper is organized as follows. Section 2 briefly introduces the related work. Section 3 illustrates the details of the proposed method. Experiment studies and discussion of results are given in Section 4. Finally, the conclusion is drawn in Section 5.

2. Related Works

2.1. Hyperspectral Reconstruction with Deep Learning Methods

Many deep learning methods have been proposed for hyperspectral reconstruction. According to their learning manner, these methods can be grossly divided into three groups: supervised learning [24,25,26,27,28,29,30], semi-supervised learning [32,33], and unsupervised learning [31]. Supervised learning methods can extract discriminative information, but are easy to overfit if the models are very complex; semi-supervised learning methods conduct supervised learning of labeled data in conjunction with unsupervised learning of unlabeled data; unsupervised learning methods can learn patterns without labeled data. In this paper, only supervised learning methods are taken into account. Supervised learning methods often use CNN models to learn feature representations. For example, inspired by the VDSR network for spatial super-resolution [37], Xiong et al. [24] utilized spectrally upsampled images to learn an end-to-end mapping. Shi et al. [25] proposed two advanced convolution neural networks (HSCNN-R and HSCNN-D) for a hyperspectral reconstruction task. With a long skip connection, a number of short residual blocks were stacked in HSCNN-R, and these short residual blocks were replaced by dense blocks with a novel fusion strategy in HSCNN-D. Both ranked in the 2nd and 1st place in NTIRE 2018 Spectral Reconstruction Challenge [38]. Zhang et al. [26] introduced dense skip connections in a deep network to extract features and map features to HSIs and designed a customized loss function to acquire the correlation among different spectra. This method outperformed several state-of-the-art methods. Zhao et al. [27] presented a 4-level hierarchical regression network (HRNet) utilizing PixelShuffle to build up inter-level interaction. This HRNet ranked 3rd in the NTIRE 2020 Spectral Reconstruction Challenge. Zhang et al. [28] employed a mixing function to adaptively determine the receptive field for each pixel. After an SE module [39] was added at the tail of each branch and each module, this pixel-aware deep function-mixture network ranked 2nd in the NTIRE 2020 Spectral Reconstruction Challenge [40]. Li et al. [29] proposed a novel network called adaptive weighted attention network (AWAN), which captured channel-wise contextual information and long-range spatial dependencies. Furthermore, this network integrated the discrepancies of RGB images and HSIs as a finer constraint for more accurate reconstruction and ranked in the 1st place in the NTIRE 2020 Spectral Reconstruction Challenge [40]. Afterwards, Li et al. [30] also proposed a hybrid 2-D-3-D deep residual attentional network (HDRAN) consisting of 2D-RANs followed by 3D-RANs to extract spatial and interband correlations simultaneously. Although these deep-learning methods can extract deep features, it is still necessary to pay more attention to some key informative features from abundant information.

2.2. Attention Mechanism

Inspired by human visual perception [41], attention mechanisms have been employed to adaptively process visual information in many applications. Wang et al. [35] proposed a non-local neural network to rescale spatial features in video classification. Hu et al. [36] proposed a squeeze-and-excitation (SE) block to model channel-wise dependencies to obtain significant improvement in performance for image classification. However, the SE block only explored first-order statistics. Therefore, Dai et al. [42] utilized second-order statistics and developed a second-order channel attention (SOCA) module for more discriminative representations. Meanwhile, considering that the traditional non-local operation was used to compute the global-level long-range dependencies, which cost a lot of memory, Dai et al. [42] proposed region-level non-local operations for image super-resolution. Anwar et al. [43] made the attention block obtain larger receptive fields by applying the Laplacian pyramid to learn the critical features at different scales for highly accurate image super-resolution. More recently, some attention mechanisms have been applied in the hyperspectral reconstruction task. Zhang et al. [40] inserted an SE block to acquire the channel-wise dependencies of a function-mixture block for hyperspectral reconstruction. Li et al. [29] proposed adaptive weighted channel attention (AWCA) module and a patch-level second-order non-local (PSNL) module to capture channel-wise and spatial contextual information, which strengthened discriminative learning of the network used to recover HSIs. Compared with the traditional non-local network, the region-level non-local and patch-level non-local modules need relatively less computational time, but the process of calculating the attention map is not efficient enough yet.

2.3. Adaptive Fusion Block

Recently, an adaptive fusion strategy has been introduced to CNN to extract more features in computer vision tasks [44,45] and remote sensing [46,47]. In [44], an adaptive weighted multi-scale module was proposed to remove some scale branches with lower contributions, and the features from nonlinear mapping module were made full use of to improve single-image super-resolution quality. In [45], adaptive weighted groups were introduced into dense links to adaptively select informative features and reduce the feature redundancy. In [46], adaptive fusion blocks were employed to preserve the features of size-varied objects from different levels, which improved the performance of remote sensing image segmentation. More recently, to better take advantage of the complementary information of spectral and spatial features, a feature fusion module was designed to adaptively adjust the voting weights of spectral features and spatial features on hyperspectral image classification results [47].

3. Methodology

In this section, we firstly introduce the network structure of DRN-DA, and then detail the densely residual attention block (DRAB), adaptive fusion block (AFB), and the adaptive multi-scale block (AMB).
Some basic nomenclature is first introduced before the proposed network is described in detail. Different convolutions are distinguished by superscript numbers. For example, C o n v 1 × 1 ( · ) denotes the function of a 1 × 1 convolutional layer, and C o n v 3 × 3 ( · ) represents the function of a 3 × 3 convolutional layer. Additionally, · and δ ( · ) denote the concatenation operation and PReLU activation function, respectively.

3.1. Network Architecture of DRN-DA

The overall architecture of DRN-DA is shown in Figure 1. DRN-DA falls into four stages, including a shallow feature extraction stage, a feature mapping stage, an adaptive fusion stage, and a reconstruction stage. Let us denote I RGB R N × 3 × H × W and I HSI R N × 31 × H × W as the input and output of DRN-DA. Here, 3 or 31 is the band number, N is the batch size, H is the height, and W is the width.
Firstly, we use two 3 × 3 convolutional layers with an activation function called a parametric rectified linear unit (PReLU) [48] between them to extract the shallow features F 0 from RGB input images
F 0 = H SF ( I RGB )
where H SF ( · ) represents the shallow feature extraction function. Then, the shallow features F 0 R N × C × H × W are fed to the feature mapping stage for higher-level feature extraction. C is the channel number of the feature map. The procedure can be described as follows:
F FM = H FM ( F 0 )
where H FM ( · ) represents the feature mapping function, which consists of M 1 DRBs and M 2 DRABs. F FM is composed of a set of feature maps as
F FM = F DRB _ 1 , , F DRB _ M 1 , F DRAB _ 1 , , F DRAB _ M 2
Then, in the adaptive fusion stage, the feature maps extracted by multiple DRBs and DRABs are adaptively fused. This procedure is expressed as
F AFB = H AFB ( F FM )
where H AFB ( · ) represents the fusion function described in Section 3.3. At the end of the adaptive fusion stage, global residual learning is introduced to keep the network stable. For the identity mapping branch, a 5 × 5 convolutional layer is utilized to further process the shallow features. The global residual learning can be formulated as
F GRL = F AFB + C o n v 5 × 5 ( F 0 )
In the final stage, the global feature representation F GRL is reconstructed via a reconstruction module as follows:
I HSI = H RC ( F GRL ) = H DRN DA ( I RGB )
where H RC ( · ) and H DRN DA ( · ) are the reconstruction module and the function of the network of DRN-DA, respectively. The reconstruction module is composed of an adaptive multi-scale block (AMB) and a 3 × 3 convolutional layer. The AMB block has three branches with multiple scale convolutions, and corresponding weights can be automatically learned to make full use of more important representations. Finally, the convolutional layer is used to compress the dimension to N × 31 × H × W , which is the same as the dimension of the ground-truth HSIs.

3.2. Densely Residual Attention Block

As shown in Figure 1, the backbone of the proposed network DRN-DA is stacked with multiple densely residual blocks (DRB) and densely residual attention blocks (DRAB). As the only difference between DRB and DRAB lies in that DRB does not incorporate attention mechanisms, while DRAB has them, here, only DRAB is described in detail. As shown in Figure 1, DRAB consists of a lightweight dense connection and local residual learning. The lightweight dense connection is inspired by DenseNet [34] to alleviate the vanishing-gradient problem and strengthen feature propagation. However, the feature maps are used as inputs into all subsequent layers in the original DenseNet, which results in the DenseNet becoming wider and wider with increasing depth. Therefore, here, a simplified densely connected network is developed to reduce the computation cost. The output of each multi-scale residual attention block (MRAB) is only used as the input into the following second MRAB. After the feature maps of two consecutive layers are concatenated, a 3 × 3 convolutional layer is added to reduce the dimension, which further lowers the width of the dense connection. The procedure of DRAB can be formulated as
x 1 = C o n v 3 × 3 ( [ f MRAB ( x 0 ) , x 0 ) ] ) x 2 = C o n v 3 × 3 ( [ f MRAB ( x 1 ) , f MRAB ( x 0 ) , x 0 ) ] ) x 3 = C o n v 3 × 3 ( [ f MRAB ( x 2 ) , f MRAB ( x 1 ) , f MRAB ( x 0 ) , x 0 ) ] )
where x 0 , x 1 , x 2 , and x 3 denote the input of the first, second, and third MRAB, and the feature map which is output from the third MRAB and processed via a convolutional layer, respectively; f MRAB ( · ) is the function of MRAB.
As shown in Figure 1, MRAB is composed of three subcomponents, namely multi-scale residual block (MRB), channel attention (CA), and dual downsampling spatial channel (DDSA). The features are processed by three branches: the first branch is stacked with three MRBs; the second branch consisting of CA is parallel to the second MRB; and the third branch consisting of DDSA is parallel to the second MRB as well. Both CA and DDSA work on the second MRB, and the feature maps are fused with concatenation and compression in the channel dimension. The procedure of MRAB is described by the following equations:
y 1 = f MRB ( y 0 ) y 2 = f MRB ( y 1 ) y 3 = f MRB ( C o n v 3 × 3 ( [ y 2 f CA ( y 1 ) , y 2 f DDSA ( y 1 ) ] ) )
where y 0 , y 1 , y 2 , and y 3 denote the input of MRAB and the outputs of the first, second, and third MRB, respectively; f MRB ( · ) , f CA ( · ) , and f DDSA ( · ) represent the functions of MRB, CA and DDSA; ⊗ and ⊕ denote the element-wise multiplication and addition, respectively.
Next, we give more details on the multi-scale residual block (MRB), channel attention (CA), and dual downsampling spatial channel (DDSA).

3.2.1. Multi-Scale Residual Block

Since it has been proven that wider features before the activation layer of the residual block can exploit the multi-level features better [49,50], a multi-scale residual block (MRB) is used in the basic building block MRAB. Different from the MRB in [50], a modified MRB is adopted. Inspired by Inception-V2 [51], the 5 × 5 convolutional layer and the 7 × 7 convolutional layer in the multi-scale convolution block from [50] are replaced by two 3 × 3 convolutional layers and three 3 × 3 convolutional layers, respectively. This strategy is adopted to reduce parameters as well as the computational time, but can make the stacked convolutions reach the same receptive fields as the convolutions in [43]. Additionally, the activation function ReLU is replaced by PReLU to introduce more nonlinearity and accelerate convergence. The MRB in this work consists of several parts: a multi-scale convolution block, a PReLU activation function, a feature fusion bottleneck layer, a 3 × 3 convolution layer, and a local residual learning block (see Figure 2).
Multi-Scale Convolution Block: The computations of the three paths in multi-scale convolution block are formulated as
s 1 = C o n v 1 , 1 3 × 3 ( s 0 ) s 2 = C o n v 2 , 2 3 × 3 ( C o n v 2 , 1 3 × 3 ( s 0 ) ) s 3 = C o n v 3 , 3 3 × 3 ( C o n v 3 , 2 3 × 3 ( C o n v 3 , 1 3 × 3 ( s 0 ) ) )
where s 0 , s 1 , s 2 , and s 3 are the input of the multi-scale convolution block and the outputs of the first, second, and third paths; C o n v 1 , 1 3 × 3 ( · ) , C o n v 2 , 1 3 × 3 ( · ) , C o n v 2 , 2 3 × 3 ( · ) , C o n v 3 , 1 3 × 3 ( · ) , C o n v 3 , 2 3 × 3 ( · ) , and C o n v 3 , 3 3 × 3 ( · ) are the 3 × 3 convolutional layer in the first path, the first and second 3 × 3 convolutional layers in the second path, and the first, second, third 3 × 3 convolutional layers in the third path.
Feature Fusion: The feature maps output from three paths are concatenated and activated. As a result, 3 × C feature maps are generated. Then, a 1 × 1 convolutional layer is used to fuse the multi-scale features and compress the number of channels. Finally, a 3 × 3 convolutional layer is employed to extract spatial-wise features. This procedure of feature fusion can be formulated as
p = C o n v 3 × 3 ( C o n v 1 × 1 ( δ ( [ s 1 , s 2 , s 3 ] ) ) )
Local Residual Learning: considering that there are multiple convolutional layers in the above architecture, local residual learning is used to enhance the feature map. The final output feature map of MRB is given by
s = p + s 0

3.2.2. Channel Attention

In previous tasks, channel-wise attention has been proven to be efficient to select significant features [29,39,40,42,43]. The channel attention in the MRAB architecture is the same as the SE layer [39] shown in Figure 3. First, the input feature U R C × H × W is compressed through a global average pooling and made into a global statistics vector q R C × 1 × 1 :
q = 1 H × W i = 1 H j = 1 W U ( i , j )
Next, two fully-connected (FC) layers are used to obtain a bottleneck. The first FC layer is a channel-reduction layer with reduction ratio r, and the second FC layer restores the number of channels. Between the two FC layers, a PRELU is used to increase the nonlinearity.
Finally, a sigmoid function is applied as a gating mechanism to acquire the channel statistics.

3.2.3. Dual Downsampling Spatial Attention

The non-local self-attention mechanism proposed by Wang [35] can capture the long-range dependencies at all positions throughout the entire image. It has been used in various computer vision applications such as video classification and image recognition, and yields great improvement. However, the standard non-local network consumes a large amount of computational memory since each position’s signal is calculated on the whole image, and the computational cost is prohibitive when the image has a large size. Though region-level non-local modules [42] and patch-level non-local modules [29] reduce computational time to some extent, these methods are not efficient enough yet. To address this problem, modified non-local attention is developed in this work to model spatial correlation. In this attention block, a dual downsampling strategy is used to reduce the channel dimension and the size of the input image as illustrated in Figure 4. Let g 0 R C × H × W be the input feature map; then, three convolutions are used to convert g 0 into three different feature maps, g 1 R ( C / s ) × H × W , g 2 R ( C / s ) × ( H / t ) × ( W / t ) , and g 3 R ( C / s ) × ( H / t ) × ( W / t ) . Concretely, the three convolutions are in the size 1 × 1 × ( C / s ) , and the latter two convolutions have a stride with t to downsample the image in size. This process can be expressed as
g 1 = C o n v 1 1 × 1 ( g 0 ) g 2 = C o n v 2 1 × 1 ( g 0 ) g 3 = C o n v 3 1 × 1 ( g 0 )
Next, the three feature maps are flattened to R ( C / s ) × H W , R ( C / s ) × ( H W / t 2 ) , and R ( C / s ) × ( H W / t 2 ) . Then, a spatial attention map is calculated through multiplying the transposed first feature map by the second feature map and normalized by a Softmax activation function as
B = τ ( g 1 T × g 2 )
where τ ( · ) denotes the Softmax function, and B is the spatial attention matrix. Subsequently, the feature map g 3 is multiplied by the transposed B to acquire the weighted feature map g 4 , which is formulated as
g 4 = g 3 × B T
At last, g 4 is reshaped to R ( C / s ) × H × W and then recovered to the same channel dimension as the input feature map g 0 by a 1 × 1 convolution. Therefore, the final output is obtained by
g 5 = C o n v 1 × 1 ( g 4 )

3.3. Adaptive Fusion Block

A traditional fusion structure integrates the hierarchical features from early layers to make full use of features from different levels [28,52]. However, most global fusion blocks treat different layers equally. This causes the network to omit more important information, and the quality of reconstructed image is reduced. Therefore, an efficient structure is developed to make the fusion block more effective. Taking the inter-layer relationships into account, in this work, an adaptive fusion block (AFB) is used to fuse the outputs of all DRBs and DRABs (see Figure 1). Each output has an independent weight that is adaptively and automatically adjusted in [0, 1] according to the training loss. The features from different layers are weighted, concatenated, and fused by a convolution. Mathematically, the procedure can be formulated as
F AFB = C o n v 3 × 3 ( [ ω DRB _ 1 F DRB _ 1 , , ω DRB _ M 1 F DRB _ M 1 , ω DRAB _ 1 F DRAB _ 1 , , ω DRAB _ M 2 F DRAB _ M 2 ] )
where ω DRB _ 1 , ω DRB _ M 1 , ω DRAB _ 1 , and ω DRAB _ M 2 are trainable weights representing weights of the first, M 1 th DRB, the first and M 2 th DRAB, respectively; C o n v 3 × 3 ( · ) reduces the channel number.

3.4. Adaptive Multi-Scale Block

For a deep learning network, a larger receptive field usually brings the network better representative capability. In this work, at the tail of the proposed network DRN-DA, an AMB with larger receptive fields than the previous multi-scale convolution block in Figure 2 is added to further improve the performance of the network as shown in Figure 1. The AMB is comprised of three kinds of scale convolutions (see Figure 5) with kernel sizes of 3 × 3, 5 × 5, and 7 × 7. After each convolution, the PReLU function is added to increase nonlinearity, and the other convolution with the same size is used to make the receptive field larger. The weight of each scale branch is automatically learned according to the training loss, which results in a trade-off between the reconstruction quality and parameters. The computational process of AMB can be formulated as
h = ω 1 × C o n v 3 × 3 ( δ ( C o n v 3 × 3 ( h 0 ) ) ) + ω 2 × C o n v 5 × 5 ( δ ( C o n v 5 × 5 ( h 0 ) ) ) + ω 3 × C o n v 7 × 7 ( δ ( C o n v 7 × 7 ( h 0 ) ) )
where h 0 and h represent the input and the output of AMB, respectively; and ω 1 , ω 2 , and ω 3 are the weights of the three branches.

4. Experiments

4.1. Settings

4.1.1. Data Sets

In this section, we perform several experiments to verify the effectiveness of the proposed network on four data sets from the NTIRE 2018 and NTIRE 2020 Spectral Reconstruction Challenges [38,40]. In each challenge, there are two data sets: the “Clean” track and the “Real World” track. The “Clean” track contains noise-free and uncompressed 8-bit RGB images obtained from a CIE 1964 color matching function, and the “Real World” track involves JPEG-compressed 8-bit RGB images created by applying an unknown camera response function to ground truth HSIs. Each track in the NTIRE 2018 challenge has 256 natural images for training and 5 + 10 additional images for validation and testing with a size of 1392 × 1300. The reflectance of HSIs in the NTIRE 2018 challenge is in the range of [0, 4095]. Each track in the NTIRE 2020 challenge has 450 images for training and 10 + 20 additional images for validation and testing with a size of 512 × 482. The reflectance of HSIs in the NTIRE 2020 challenge has been normalized to be in [0, 1]. All the HSIs have 31 bands from 400 nm to 700 nm, with a 10 nm step. Since the official testing HSIs in both challenges are confidential, in this work, the official validation HSIs are used for testing, which is the same as the work [30].

4.1.2. Experimental Configuration

In the proposed DRN-DA, we adopted 2 DRBs and 2 DRABs, namely, M 1 = M 2 = 2. In the intermediate layers, the number of channel dimension C was 64. The channel reduction ratio r of the CA module was 16, and the channel reduction ratio s and the size reduction ratio t of DDSA module were set to be 8 and 4, respectively.
In the training stage, all the training images were resized to be 64 × 64. The model was trained using a batch size of 32 with the Adam optimizer [53] by setting β 1 = 0.9 , β 2 = 0.999 , and ε = 10 8 . The learning rate was initialized as 10 4 , and the polynomial function was set as the decay policy with power = 1.5 . The training process was terminated at the 100th epoch. The model was trained using a PyTorch framework and a NVIDIA RTX 3090 GPU.

4.1.3. Quantitative Metrics

As in the NTIRE 2018 and NTIRE 2020 challenges [38,40], mean relative absolute error (MRAE) is chosen as the quantitative metric of the proposed method. MRAE is defined as
MRAE = 1 N i = 1 N r i r i r i
where r i and r i denote the ground-truth and reconstructed spectral vectors of the ith pixel, respectively; N is the total number of pixels.
MRAE represents the spatial distance between the ground-truth and reconstructed spectral curves. In general, a smaller MRAE indicates better performance. Moreover, the peak signal-to-noise ratio (PSNR), structural similarity index (SSIM) [54], and spectral angle mapper index (SAM) [55] are used as complementary measures. PSNR and SSIM measure the similarity between the ground truth and the reconstructed spectra. The higher the measures, the better the quality. SAM denotes the angle between the two spectral vectors. The lower the measure, the better the quality.

4.2. Comparisons with State-of-the-Art Methods

As comparisons, several state-of-the-art methods were selected to be compared with the proposed DRN-DA. These methods included Arad [22], HSCNN-R [25], HSCNN-D [25], PDFM [28,40], AWAN [29], and HDRAN [30]. Arad is a method based on sparse coding, and except for HDRAN, the other methods are the top two methods in the two challenges. In addition, to further improve the performance of DRN-DA and make the results more convincing, an ensemble strategy was introduced; the ensemble DRN-DA is denoted as DRN-DA+. Specifically, DRN-DA was trained three times without adjusting any parameters, and the testing results were averaged into the final results, which can avoid the influence of random factors on the results. For a fair comparison, other CNN-based methods use ensemble strategies as well. As the ensemble strategies in the original papers [25,29], HSCNN-R, HSCNN-D, and AWAN with different depths and widths were trained and denoted as HSCNN-R+, HSCNN-D+, and AWAN+. Similar to DRN-DA+, an ensemble strategy was applied to PDFN and HDRAN, and the corresponding models were denoted as PDFN+ and HDRAN+.

4.2.1. Quantitative Evaluation

The results of the hyperspectral reconstruction are summarized in Table 1 and Table 2. For the “Real World” tracks, a self-ensemble strategy used in [29] was adopted, which means that the RGB testing images were flipped up/down and fed to the network. Then, the mirror output and the original output were averaged into the final results. It is noted that the Arad method is only suitable for the “Clean” tracks since the camera response function generating the “Real World” tracks is unknown. Except for HSCNN-R, HSCNN-D on the NTIRE 2018 data sets and AWAN on the NTIRE 2020 “Clean” track, the other networks learned from scratch.
In Table 1 and Table 2, the best and second-best results are bold and underlined, respectively. As can be seen, the proposed DRN-DA performs better than previous methods. Specifically, compared with the second-best method, DRN-DA brings a decrease in MRAE of 6.19% and 1.43% on the NTIRE 2018 “Clean” track and “Real World” track; it brings a decrease in MRAE of 6.85% and 5.30% on the NTIRE 2020 “Clean” track and “Real World” track. In addition, for other measures, DRN-DA also performs better than or equally well as the second-best method. Overall, the CNN-based methods perform much better than the Arad method.
In Table 3 and Table 4, the best and second-best results are bold and underlined, respectively. It can be seen that the proposed DRN-DA+ also outperforms other methods on the four data sets in terms of MRAE, PSNR, and SAM, though it performs equally with HDRAN+ on the NTIRE 2018 “Real World” track in terms of SSIM. Specifically, compared with the second-best method, it produces a reduction in MRAE of 4.59% and 1.84% on the NTIRE 2018 “Clean” track and “Real World” track and a reduction in MRAE of 5.13% and 4.06% on the NTIRE 2020 “Clean” track and “Real World” track.

4.2.2. Visual Evaluation

In order to compare the reconstruction quality of different methods visually, the error maps of four images on five sampled bands (450 nm, 500 nm, 550 nm, 600 nm, and 650 nm) are shown in Figure 6, Figure 7, Figure 8 and Figure 9. The first and last rows show the ground-truth HSIs and the error maps reconstructed by the proposed DRN-DA, and the middle rows show the error maps reconstructed by other benchmark methods. These error maps are the heat maps of MRAE between the ground truth and the reconstructed spectra. As seen from Figure 6, Figure 7, Figure 8 and Figure 9, for the example images from the NTIRE 2018 “Clean” track, and the NTIRE 2020 “Clean” track and “Real World” track, the superiority of the proposed DRN-DA is more significant at the long wavelength 650 nm; however, for the example image from the NTIRE 2018 “Real World” track, the differences between the proposed DRN-DA and other benchmark methods are larger at the long wavelength 600 nm. On the whole, the proposed method generates smaller spectral errors.
Moreover, two samples from each example image are selected (see Figure 10), and the corresponding spectra recovered by different methods are plotted in Figure 11, Figure 12, Figure 13 and Figure 14, where the black and red curves denote the ground-truth and recovered spectra by the proposed method, and the light blue, green, blue, magenta, cyan, and orange curves represent the recovered spectra from Arad, HSCNN-R, HSCNN-D, PDFN, AWAN, and HDRAN, respectively. It can be observed that the recovered spectra from the proposed DRN-DA are closer to the ground-truth spectra.

4.3. Ablation Study for Different Modules

In this section, an ablation study is carried out on NTIRE 2020 data sets to verify the effects of different ingredients, including CA, DDSA, AMB, and AFB in the proposed network. The results in MRAE are reported in Table 5. E a is a baseline network, which only contains ordinary convolutions. Based on E a , CA, DDSA, AMB, and AFB are added in succession to the baseline network to obtain E b , E c , E d , and E e . The gradual reduction of MRAEs on the two data sets demonstrates that these blocks do work for spectral reconstruction.

4.4. Effectiveness of Multi-Scale Convolution Scheme

To validate the effectiveness of the multi-scale convolution scheme, we experimentally analyzed the effects of multi-scale convolutions based on the NTIRE 2020 data sets. As shown in Table 6, the performance increases with the use of multi-scale convolutions. For the NTIRE 2020 “Clean” track and “Real World” track, the network with multi-scale convolutions gains over the network with 3 × 3 convolutions with 3.86% and 4.14% in terms of MRAE, respectively.

5. Conclusions

In this paper, we propose a deep neural network (DRN-DA) for hyperspectral recovery from a single RGB image. The basic block of DRN-DA adopts a simplified densely connected structure, which can reuse features and avoid the basic block from becoming too wide. Different from the original nonlocal module, a dual downsampling strategy is applied to learn the long-range contextual features with much less computation and memory consumption. The dual downsampling spatial attention module, together with a channel attention module, obtains spatial-wise and channel-wise features in a parallel manner to enhance the discriminative ability of the network. Moreover, the adaptive fusion strategy is utilized to fuse multiple-layer features to explore the correlated information from inter-layers via dynamically computing the weight of each layer. Extensive experiments over four data sets demonstrate that the proposed method provides better performance compared with several state-of-the-art methods. However, the efficiency of the network is not taken into account. In the future, we will work on a lightweight hyperspectral reconstruction network to save runtime. Moreover, we hope that this work can be applied to real-world applications.

Author Contributions

L.W. conceived the study; L.W. trained the model, analyzed the data, and wrote the paper; A.S. and J.Y.H. reviewed and improved the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by NTNU-CSC Joint Scholarship.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pelagotti, A.; Del Mastio, A.; De Rosa, A.; Piva, A. Multispectral imaging of paintings. IEEE Signal Process. Mag. 2008, 25, 27–36. [Google Scholar] [CrossRef]
  2. Chane, C.S.; Mansouri, A.; Marzani, F.S.; Boochs, F. Integration of 3D and multispectral data for cultural heritage applications: Survey and perspectives. Image Vis. Comput. 2013, 31, 91–102. [Google Scholar] [CrossRef] [Green Version]
  3. Nishidate, I.; Maeda, T.; Niizeki, K.; Aizu, Y. Estimation of melanin and hemoglobin using spectral reflectance images reconstructed from a digital RGB image by the Wiener estimation method. Sensors 2013, 13, 7902–7915. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Yuen, P.W.; Richardson, M. An introduction to hyperspectral imaging and its application for security, surveillance and target acquisition. Imaging Sci. J. 2010, 58, 241–253. [Google Scholar] [CrossRef]
  5. Chao, K.; Yang, C.C.; Chen, Y.; Kim, M.; Chan, D. Hyperspectral-multispectral line-scan imaging system for automated poultry carcass inspection applications for food safety. Poult. Sci. 2007, 86, 2450–2460. [Google Scholar] [CrossRef]
  6. Valero, E.M.; Hu, Y.; Hernández-Andrés, J.; Eckhard, T.; Nieves, J.L.; Romero, J.; Schnitzlein, M.; Nowack, D. Comparative performance analysis of spectral estimation algorithms and computational optimization of a multispectral imaging system for print inspection. Color Res. Appl. 2014, 39, 16–27. [Google Scholar] [CrossRef]
  7. Pan, Z.; Healey, G.; Prasad, M.; Tromberg, B. Face recognition in hyperspectral images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1552–1560. [Google Scholar]
  8. Van Nguyen, H.; Banerjee, A.; Chellappa, R. Tracking via object reflectance using a hyperspectral video camera. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, San Francisco, CA, USA, 13–18 June 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 44–51. [Google Scholar]
  9. Li, J.; Du, Q.; Li, Y.; Li, W. Hyperspectral image classification with imbalanced data based on orthogonal complement subspace projection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3838–3851. [Google Scholar] [CrossRef]
  10. Wu, F.; Duan, J.; Chen, S.; Ye, Y.; Ai, P.; Yang, Z. Multi-target recognition of bananas and automatic positioning for the inflorescence axis cutting point. Front. Plant Sci. 2021, 12, 705021. [Google Scholar] [CrossRef]
  11. Tang, Y.; Zhu, M.; Chen, Z.; Wu, C.; Chen, B.; Li, C.; Li, L. Seismic performance evaluation of recycled aggregate concrete-filled steel tubular columns with field strain detected via a novel mark-free vision method. Structures 2022, 37, 426–441. [Google Scholar] [CrossRef]
  12. Green, R.O.; Eastwood, M.L.; Sarture, C.M.; Chrien, T.G.; Aronsson, M.; Chippendale, B.J.; Faust, J.A.; Pavri, B.E.; Chovit, C.J.; Solis, M.; et al. Imaging spectroscopy and the airborne visible/infrared imaging spectrometer (AVIRIS). Remote Sens. Environ. 1998, 65, 227–248. [Google Scholar] [CrossRef]
  13. Hearn, D.; Digenis, C.; Lencioni, D.; Mendenhall, J.; Evans, J.; Welsh, R. EO-1 advanced land imager overview and spatial performance. In Proceedings of the IGARSS 2001, Scanning the Present and Resolving the Future, IEEE 2001 International Geoscience and Remote Sensing Symposium (Cat. No. 01CH37217), Sydney, Australia, 9–13 July 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 2, pp. 897–900. [Google Scholar]
  14. Hagen, N.A.; Gao, L.S.; Tkaczyk, T.S.; Kester, R.T. Snapshot advantage: A review of the light collection improvement for parallel high-dimensional measurement systems. Opt. Eng. 2012, 51, 111702. [Google Scholar] [CrossRef] [PubMed]
  15. Li, Q.; He, X.; Wang, Y.; Liu, H.; Xu, D.; Guo, F. Review of spectral imaging technology in biomedical engineering: Achievements and challenges. J. Biomed. Opt. 2013, 18, 100901. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Shen, H.L.; Cai, P.Q.; Shao, S.J.; Xin, J.H. Reflectance reconstruction for multispectral imaging by adaptive Wiener estimation. Opt. Express 2007, 15, 15545–15554. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Zhang, W.F.; Tang, G.; Dai, D.Q.; Nehorai, A. Estimation of reflectance from camera responses by the regularized local linear model. Opt. Lett. 2011, 36, 3933–3935. [Google Scholar] [CrossRef]
  18. Wang, L.; Wan, X.; Xiao, G.; Liang, J. Sequential adaptive estimation for spectral reflectance based on camera responses. Opt. Express 2020, 28, 25830–25842. [Google Scholar] [CrossRef]
  19. Nguyen, R.M.; Prasad, D.K.; Brown, M.S. Training-based spectral reconstruction from a single RGB image. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 186–201. [Google Scholar]
  20. Aeschbacher, J.; Wu, J.; Timofte, R. In defense of shallow learned spectral reconstruction from rgb images. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 471–479. [Google Scholar]
  21. Robles-Kelly, A. Single image spectral reconstruction for multimedia applications. In Proceedings of the 23rd ACM international conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 251–260. [Google Scholar]
  22. Arad, B.; Ben-Shahar, O. Sparse recovery of hyperspectral signal from natural RGB images. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 19–34. [Google Scholar]
  23. Jia, Y.; Zheng, Y.; Gu, L.; Subpa-Asa, A.; Lam, A.; Sato, Y.; Sato, I. From RGB to spectrum for natural scenes via manifold-based mapping. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4705–4713. [Google Scholar]
  24. Xiong, Z.; Shi, Z.; Li, H.; Wang, L.; Liu, D.; Wu, F. Hscnn: Cnn-based hyperspectral image recovery from spectrally undersampled projections. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 518–525. [Google Scholar]
  25. Shi, Z.; Chen, C.; Xiong, Z.; Liu, D.; Wu, F.H. Advanced cnn-based hyperspectral recovery from rgb images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 18–22. [Google Scholar]
  26. Zhang, J.; Sun, Y.; Chen, J.; Yang, D.; Liang, R. Deep-learning-based hyperspectral recovery from a single RGB image. Opt. Lett. 2020, 45, 5676–5679. [Google Scholar] [CrossRef]
  27. Zhao, Y.; Po, L.M.; Yan, Q.; Liu, W.; Lin, T. Hierarchical regression network for spectral reconstruction from RGB images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 422–423. [Google Scholar]
  28. Zhang, L.; Lang, Z.; Wang, P.; Wei, W.; Liao, S.; Shao, L.; Zhang, Y. Pixel-aware deep function-mixture network for spectral super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12821–12828. [Google Scholar]
  29. Li, J.; Wu, C.; Song, R.; Li, Y.; Liu, F. Adaptive weighted attention network with camera spectral sensitivity prior for spectral reconstruction from RGB images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 462–463. [Google Scholar]
  30. Li, J.; Wu, C.; Song, R.; Xie, W.; Ge, C.; Li, B.; Li, Y. Hybrid 2-D–3-D Deep Residual Attentional Network With Structure Tensor Constraints for Spectral Super-Resolution of RGB Images. IEEE Trans. Geosci. Remote Sens. 2020, 59, 2321–2335. [Google Scholar] [CrossRef]
  31. Zhang, J.; Zhao, D.; Chen, J.; Sun, Y.; Yang, D.; Liang, R. Unsupervised learning for hyperspectral recovery based on a single RGB image. Opt. Lett. 2021, 46, 3977–3980. [Google Scholar] [CrossRef]
  32. Lore, K.G.; Reddy, K.K.; Giering, M.; Bernal, E.A. Generative adversarial networks for spectral super-resolution and bidirectional rgb-to-multispectral mapping. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 926–933. [Google Scholar]
  33. Liu, P.; Zhao, H. Adversarial Networks for Scale Feature-Attention Spectral Image Reconstruction from a Single RGB. Sensors 2020, 20, 2426. [Google Scholar] [CrossRef]
  34. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  35. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  37. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  38. Arad, B.; Ben-Shahar, O.; Timofte, R. Ntire 2018 challenge on spectral reconstruction from rgb images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 929–938. [Google Scholar]
  39. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  40. Arad, B.; Timofte, R.; Ben-Shahar, O.; Lin, Y.T.; Finlayson, G.D. Ntire 2020 challenge on spectral reconstruction from an rgb image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 446–447. [Google Scholar]
  41. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, A. Recurrent models of visual attention. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 2204–2212. [Google Scholar]
  42. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  43. Anwar, S.; Barnes, N. Densely residual laplacian super-resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1192–1204. [Google Scholar] [CrossRef] [PubMed]
  44. Wang, C.; Li, Z.; Shi, J. Lightweight image super-resolution with adaptive weighted learning network. arXiv 2019, arXiv:1904.02358. [Google Scholar]
  45. Chen, L.; Yang, X.; Jeon, G.; Anisetti, M.; Liu, K. A trusted medical image super-resolution method based on feedback adaptive weighted dense network. Artif. Intell. Med. 2020, 106, 101857. [Google Scholar] [CrossRef] [PubMed]
  46. Liu, R.; Mi, L.; Chen, Z. AFNet: Adaptive fusion network for remote sensing image semantic segmentation. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7871–7886. [Google Scholar] [CrossRef]
  47. Gao, H.; Chen, Z.; Xu, F. Adaptive spectral-spatial feature fusion network for hyperspectral image classification using limited training samples. Int. J. Appl. Earth Obs. Geoinf. 2022, 107, 102687. [Google Scholar] [CrossRef]
  48. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  49. Yu, J.; Fan, Y.; Yang, J.; Xu, N.; Wang, Z.; Wang, X.; Huang, T. Wide activation for efficient and accurate image super-resolution. arXiv 2018, arXiv:1808.08718. [Google Scholar]
  50. Zhang, S.; Yuan, Q.; Li, J.; Sun, J.; Zhang, X. Scene-adaptive remote sensing image super-resolution using a multiscale attention network. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4764–4779. [Google Scholar] [CrossRef]
  51. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning (PMLR), Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  52. Jiang, K.; Wang, Z.; Yi, P.; Jiang, J. Hierarchical dense recursive network for image super-resolution. Pattern Recognit. 2020, 107, 107475. [Google Scholar] [CrossRef]
  53. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  54. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  55. Yuhas, R.H.; Boardman, J.W.; Goetz, A.F. Determination of semi-arid landscape endmembers and seasonal trends using convex geometry spectral unmixing techniques. In Proceedings of the JPL, Summaries of the 4th Annual JPL Airborne Geoscience Workshop, Washington, DC, USA, 25–29 October 1993; Volume 1. [Google Scholar]
Figure 1. Network architecture of DRN-DA.
Figure 1. Network architecture of DRN-DA.
Remotesensing 14 03128 g001
Figure 2. Multi-scale residual block (MRB).
Figure 2. Multi-scale residual block (MRB).
Remotesensing 14 03128 g002
Figure 3. Channel attention (CA) module.
Figure 3. Channel attention (CA) module.
Remotesensing 14 03128 g003
Figure 4. Dual downsampling spatial attention (DDSA) module.
Figure 4. Dual downsampling spatial attention (DDSA) module.
Remotesensing 14 03128 g004
Figure 5. Adaptive multi-scale block (AMB).
Figure 5. Adaptive multi-scale block (AMB).
Remotesensing 14 03128 g005
Figure 6. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2018 “Clean” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Figure 6. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2018 “Clean” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Remotesensing 14 03128 g006
Figure 7. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2018 “Real World” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Figure 7. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2018 “Real World” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Remotesensing 14 03128 g007
Figure 8. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2020 “Clean” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Figure 8. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2020 “Clean” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Remotesensing 14 03128 g008
Figure 9. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2020 “Real World” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Figure 9. Visual hyperspectral reconstruction results and visual comparisons of five selected bands for hyperspectral reconstruction error maps from one NTIRE 2020 “Real World” RGB image. The significant differences are shown in the red frames. Please zoom in for better view.
Remotesensing 14 03128 g009
Figure 10. Two selected samples in each example image. (a) NTIRE 2018 “Clean” track. (b) NTIRE 2018 “Real Word” track. (c) NTIRE 2020 “Clean” track. (d) NTIRE 2020 “Real Word” track.
Figure 10. Two selected samples in each example image. (a) NTIRE 2018 “Clean” track. (b) NTIRE 2018 “Real Word” track. (c) NTIRE 2020 “Clean” track. (d) NTIRE 2020 “Real Word” track.
Remotesensing 14 03128 g010
Figure 11. Spectral reflectance of selected samples from NTIRE 2018 “Clean” track. (a) Sample 1. (b) Sample 2.
Figure 11. Spectral reflectance of selected samples from NTIRE 2018 “Clean” track. (a) Sample 1. (b) Sample 2.
Remotesensing 14 03128 g011
Figure 12. Spectral reflectance of selected samples from the NTIRE 2018 “Real World” track: (a) Sample 1. (b) Sample 2.
Figure 12. Spectral reflectance of selected samples from the NTIRE 2018 “Real World” track: (a) Sample 1. (b) Sample 2.
Remotesensing 14 03128 g012
Figure 13. Spectral reflectance of selected samples from the NTIRE 2020 “Clean” track. (a) Sample 1. (b) Sample 2.
Figure 13. Spectral reflectance of selected samples from the NTIRE 2020 “Clean” track. (a) Sample 1. (b) Sample 2.
Remotesensing 14 03128 g013
Figure 14. Spectral reflectance of selected samples from the NTIRE 2020 “Real World” track. (a) Sample 1. (b) Sample 2.
Figure 14. Spectral reflectance of selected samples from the NTIRE 2020 “Real World” track. (a) Sample 1. (b) Sample 2.
Remotesensing 14 03128 g014
Table 1. Comparison between the proposed method and other state-of-the-art methods without the ensemble strategy based on the NTIRE 2018 data sets.
Table 1. Comparison between the proposed method and other state-of-the-art methods without the ensemble strategy based on the NTIRE 2018 data sets.
MethodNTIRE 2018
CleanReal World
MRAEPSNRSAMSSIMMRAEPSNRSAMSSIM
Arad0.074634.48484.80860.9507----
HSCNN-R0.014049.95681.04320.99880.030345.22281.61760.9952
HSCNN-D0.013550.48730.99290.99880.029345.38761.59440.9953
PDFN0.012451.51430.90130.99900.028845.71871.51970.9956
AWAN0.011552.25880.80220.99930.028745.73251.50350.9956
HDRAN0.011352.19240.80380.99920.027945.81221.45780.9957
DRN-DA0.010652.92490.74780.99940.027545.92951.45010.9957
Table 2. Comparison between the proposed method and other state-of-the-art methods without the ensemble strategy based on the NTIRE 2020 data sets.
Table 2. Comparison between the proposed method and other state-of-the-art methods without the ensemble strategy based on the NTIRE 2020 data sets.
MethodNTIRE 2020
CleanReal World
MRAEPSNRSAMSSIMMRAEPSNRSAMSSIM
Arad0.088630.05836.31120.9366----
HSCNN-R0.038938.48372.68340.99050.068735.81323.59550.9777
HSCNN-D0.038339.04262.63300.99150.070235.85283.56330.9760
PDFN0.036240.24932.40510.99360.067435.93533.41060.9781
AWAN0.032140.77672.21080.99400.066636.28593.37930.9793
HDRAN0.033840.35832.27060.99410.066036.22873.28870.9777
DRN-DA0.029941.38522.05160.99520.062536.88413.09450.9814
Table 3. Comparison between the proposed method and other state-of-the-art methods with the ensemble strategy based on the NTIRE 2018 data sets.
Table 3. Comparison between the proposed method and other state-of-the-art methods with the ensemble strategy based on the NTIRE 2018 data sets.
MethodNTIRE 2018
CleanReal World
MRAEPSNRSAMSSIMMRAEPSNRSAMSSIM
HSCNN-R+0.013550.45260.99190.99890.029745.36891.58890.9953
HSCNN-D+0.013250.63990.98290.99880.028745.59241.54690.9955
PDMN+0.012051.66190.87420.99910.028345.82061.50660.9956
AWAN+0.011252.45660.78230.99930.028245.81961.47880.9957
HDRAN+0.010952.47580.77120.99930.027745.90151.44810.9958
DRN-DA+0.010453.06000.73580.99940.027245.98641.43110.9958
Table 4. Comparison between the proposed method and other state-of-the-art methods with the ensemble strategy based on the NTIRE 2020 data sets.
Table 4. Comparison between the proposed method and other state-of-the-art methods with the ensemble strategy based on the NTIRE 2020 data sets.
MethodNTIRE 2020
CleanReal World
MRAEPSNRSAMSSIMMRAEPSNRSAMSSIM
HSCNN-R+0.037239.23372.55440.99200.067336.04953.41310.9785
HSCNN-D+0.037739.16972.60120.99180.069636.01323.50280.9767
PDMN+0.033140.49812.21440.99850.066036.17703.31030.9789
AWAN+0.031240.99872.15520.99430.064636.27573.23280.9795
HDRAN+0.033740.47862.26110.99370.064036.45803.20300.9793
DRN-DA+0.029641.46522.02790.99540.061436.80293.03510.9811
Table 5. Effects of different ingredients in the proposed network.
Table 5. Effects of different ingredients in the proposed network.
E a E b E c E d E e
CA
DDSA
AMB
AFB
Clean0.03710.03480.03430.03380.0299
Real World0.06700.06520.06490.06370.0625
Table 6. Effects of multi-scale convolutions in the proposed network.
Table 6. Effects of multi-scale convolutions in the proposed network.
E f E e
Multi-scale convolutions
Clean0.03110.0299
Real World0.06520.0625
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, L.; Sole, A.; Hardeberg, J.Y. Densely Residual Network with Dual Attention for Hyperspectral Reconstruction from RGB Images. Remote Sens. 2022, 14, 3128. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14133128

AMA Style

Wang L, Sole A, Hardeberg JY. Densely Residual Network with Dual Attention for Hyperspectral Reconstruction from RGB Images. Remote Sensing. 2022; 14(13):3128. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14133128

Chicago/Turabian Style

Wang, Lixia, Aditya Sole, and Jon Yngve Hardeberg. 2022. "Densely Residual Network with Dual Attention for Hyperspectral Reconstruction from RGB Images" Remote Sensing 14, no. 13: 3128. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14133128

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop