Next Article in Journal
Distribution and Characteristics of Damming Landslides Triggered by 1920 M~8 Haiyuan Earthquake, NW China
Previous Article in Journal
Cross-Modal Feature Representation Learning and Label Graph Mining in a Residual Multi-Attentional CNN-LSTM Network for Multi-Label Aerial Scene Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TE-SAGAN: An Improved Generative Adversarial Network for Remote Sensing Super-Resolution Images

1
School of Computer Science, China University of Geosciences, Wuhan 430074, China
2
National Engineering Research Center of Geographic Information System, China University of Geosciences, Wuhan 430074, China
3
Key Laboratory of Urban Land Resources Monitoring and Simulation, Ministry of Natural Resources, Shenzhen 518034, China
4
School of Geography and Information Engineering, China University of Geosciences, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(10), 2425; https://doi.org/10.3390/rs14102425
Submission received: 30 March 2022 / Revised: 13 May 2022 / Accepted: 17 May 2022 / Published: 18 May 2022

Abstract

:
Resolution is a comprehensive reflection and evaluation index for the visual quality of remote sensing images. Super-resolution processing has been widely applied for extracting information from remote sensing images. Recently, deep learning methods have found increasing application in the super-resolution processing of remote sensing images. However, issues such as blurry object edges and existing artifacts persist. To overcome these issues, this study proposes an improved generative adversarial network with self-attention and texture enhancement (TE-SAGAN) for remote sensing super-resolution images. We first designed an improved generator based on the residual dense block with a self-attention mechanism and weight normalization. The generator gains the feature extraction capability and enhances the training model stability to improve edge contour and texture. Subsequently, a joint loss, which is a combination of L1-norm, perceptual, and texture losses, is designed to optimize the training process and remove artifacts. The L1-norm loss is designed to ensure the consistency of low-frequency pixels; perceptual loss is used to entrench medium- and high-frequency details; and texture loss provides the local features for the super-resolution process. The results of experiments using a publicly available dataset (UC Merced Land Use dataset) and our dataset show that the proposed TE-SAGAN yields clear edges and textures in the super-resolution reconstruction of remote sensing images.

Graphical Abstract

1. Introduction

The development of optical remote sensing technology has enabled researchers to obtain large-area, multi-temporal Earth observation images for social construction with ease. Nevertheless, optical satellite image quality suffers from the sensor imaging technology, limiting the quality and application of remote sensing images. Low-resolution remote sensing images bring trouble to ground object interpretations. Therefore, super-resolution reconstruction of remote sensing images has become an indispensable preprocessing step. Improving the spatial resolution of remote sensing images promotes clearer object boundaries and contours in remote sensing images, which provides clear data for subsequent applications of extracting features, such as the recognition roads and buildings by semantic segmentation methods [1,2,3] or change detection by extracting the object boundary [4,5]. Therefore, the super-resolution processing of remote sensing images has become an indispensable preprocessing process. Many researchers have utilized super-resolution techniques to effectively and accurately obtain images [6,7,8,9,10]. Super-resolution (SR) imaging is a technology that uses single or multiple low-resolution (LR) images to obtain high-resolution (HR) images using existing image acquisition devices. It has been widely applied since the 1980s, and Huang et al. [11] creatively proposed a method for reconstructing HR images based on the frequency field of multi-frame sequential LR images. Since then, SR processing of remote sensing images has become an active research topic, and numerous pertinent algorithms have been designed. At present, researchers have divided single-image super-resolution (SISR) processing methods into interpolation-, reconstruction-, and learning-based methods.
The interpolation-based method estimates the unknown pixel value via interpolation in the designed linear function according to the known pixels, such as nearest-neighbor interpolation [12], Lanczos filtering [13], or bicubic filtering [14]. Although these methods can achieve rapid SR image processing, filling the pixels with low-frequency information leads to shortcomings in the overall semantic features. Consequently, the super-resolution images may lose information of some details or boundaries may not be clear. Reconstruction-based methods aim to establish an appropriate prior constraint or registration model between LR and HR images, such as projection onto convex sets [15,16,17], iterative back projection [18], or maximum a posteriori probability method [19,20,21]. The core objective of reconstruction-based methods is to build a reasonable prior observation model. However, SR imaging is a highly underdetermined infinite-solution problem in mathematical theory. Even using the optimal parameters, the transformation performance of the model is weak in diverse tasks. Owing to this problem, the model requires considerable resources to accomplish repeatable and complicated calculations.
In recent years, with the development of machine learning frameworks, academic interest in learning-based methods has increased, particularly deep learning driven by big data [22]. Convolutional neural networks are a representative adaptive learning method widely used for the reconstruction of low- and high-frequency images and are less complex than traditional manual interventions. Researchers have thus expanded and applied convolutional neural networks to SR image tasks [23,24,25,26,27,28]. However, both early shallow networks with 3–5 layers [29,30] and deep super-resolution networks [31,32,33] tend to produce low-frequency results.
To address these issues, researchers have explored algorithms that yield results consistent with human observations and sought alternatives capable of obtaining high-frequency images. Ledig et al. [34] proposed an improved generative adversarial network (GAN) model, named SRGAN, which benefits from content and perceptual losses and exerts good performance. Based on SRGAN, the Residual-in-Residual Dense Block (RRDB) was introduced, which further improved the recovered image texture [35]. In the last years, many ESRGAN-based algorithms [36,37,38] have been improved. ESRGAN+ [36] replaces ESRGAN’s RRDB with RRDRB; Real-ESRGAN replaces the VGG-Net type discriminator in the original ESRGAN with a U-Net type discriminator [37], which has a good improvement in anime images and videos. However, these network parameters are too large to fit training conditions. When these methods are used to reconstruct remote sensing images, the texture details are still unclear, and some artifacts remain in the produced images.
The primary reasons for the abovementioned problems can be summarized as follows: (1) Owing to their peculiar characteristics compared to conventional images of the same image dimensions, remote sensing images contain abundant information, such as external contours and internal textures, that characterize the spatial relationships of ground objects, thereby making the content appear crowded. Therefore, algorithms that can accurately recover and characterize contours and internal textures are necessary. (2) Most existing methods do not consider global spatial connectivity and refining texture in the algorithm design. Owing to the local perception characteristic of convolutional neural networks, global feature information in large-scale remote scenes is ignored. Thus, the global remote sensing information is neglected and underutilized.
To resolve these concerns, we present TE-SAGAN, a GAN-based texture-enhanced SR processing network for remote sensing images, which sharpens SR image details while maintaining clear visual edges. The self-attention mechanism (SAM) is designed to learn global information from remote sensing images. Furthermore, the RRDB was introduced for image depth extraction. Simultaneously, a SAM with three submodules was used to obtain different levels of attention, where relevant information (such as edge contours) can be engaged and received dynamically. Weight normalization (WN) [39] was also introduced to TE-SAGAN to remove artifacts and increase the training network stability. The calculated texture loss by the Gram matrix was designed to assess the difference between the original HR and SR images in high-frequency features, which can help improve the internal details of the reconstructed image.
In conclusion, by the self-attention module and texture loss units, this work provides insight into refining the image edges and textures. The major contributions are as follows: (1) designing the self-attention mechanism module for extracting the global information, which can decrease the limitation of the local receptive field and increase the super-resolution network accuracy; (2) a weight normalization layer is introduced instead of batch normalization, which can reduce the interference of mini-batch data on reconstructed image quality and improve the image artifact; (3) a joint loss is designed combining the content loss, perceptual loss, adversarial loss, and texture loss, which take into account the consistency of the pixel content and texture details and promote the super resolution work to achieve a characteristic remote sensing image.

2. Method

2.1. Principle of the Proposed Method

The designed network framework proposed in this study is based on GAN [40] and comprises two parts: a generator network and a discriminator network. The generator network consists of feed-forward neural networks to form the feature extraction module, which is expressed by the function G ( I L R ; θ G ) , where θ G denotes the weights and biases ( w L , b L ) of the neural network layer. The fully trained network was generated as follows:
I S R = G θ G ( I L R ) ,
where I S R is the super-resolution image, and I L R is the corresponding low-resolution image. I S R is as close as possible to the sample distribution of I H R , and the parameter optimization of the training network must satisfy the following rule:
θ G = arg min 1 N n = 1 N L ( I S R , I H R ) ,
where L ( I S R , I H R ) denotes the reconstruction error of a pair of data group I n ( I S R , I H R ) .
The parameters of the discriminator D were defined accordingly as θ D . Their adversarial training is essentially a minimum–maximum optimization. We converted the training parameters corresponding to the LR-HR remote sensing image processing as follows:
min G   max D V D θ D , G θ G = E I H R ~ P d a t a I H R log D θ G I H R + E I L R ~ p G I L R log 1 D G θ G I L R ,
The basic idea expressed in the above equation is that the generator network G should be well-trained under the GAN (Figure 1). The discriminator network D has to distinguish the true and false states of the input remote sensing images. For the inputted I L R , the generator in the latter term emulates the G θ G ( I L R ) sample from the true sample P d a t a ( x ) , thereby minimizing the probability that the discriminator distinguishes G θ G ( I L R ) as a fake sample. The network parameters are updated iteratively under strict monitoring conditions (e.g., loss constraints of mean square error [MSE] and mean absolute error [MAE]). The ability of the generator to fit the simulation is continuously enhanced, making it difficult for the discriminator to distinguish the reliability of the input samples. The well-trained model can predict the potential frequency distribution of the samples and generate new high-quality data sources.

2.2. Generator of TE-SAGAN

Owing to the limitation of “local perception” in conventional convolution modules, long-distance global information cannot be used for HR remote sensing images. To extract more edge and global information from remote sensing images, the generator of TE-SAGAN was designed with a low-frequency feature extractor, deep residual dense block, self-attention module, and WN (Figure 2). The self-attention module was designed to further enhance the ability of the network to extract global object features. Compared with the common batch normalization (BN) usually used in neural networks, the WN introduced in this study can stabilize network training and remove image artifacts.
In the generator network, the low-frequency feature extractor is a convolution layer with 3 × 3 kernels, following a WN layer. RRDB is an excellent nonlinear feature extraction module that has an interwoven ensemble with dense connectivity and residual skipping. RRDB can learn deep semantic features while improving low-level feature extraction. Here, we gathered the feature-flow output from RRDB with the previous residual layer, merging them into the SAM module to generate global contextual outlines. To achieve the SR processing for remote sensing images with upscaling factors (4×), these extracted features were recovered using an up-sample layer with nearest-neighbor interpolation. In this process, in addition to keeping 23 RRDBs with the residual scaling factor β = 0.2 ( 0 < β < 1 ) and a self-attention mechanism module, each regular convolution layer was followed by WN (Figure 3).

2.2.1. Self-Attention Mechanism

Traditional convolution operations associate regional pixels with neighboring pixels using a constant convolution kernel size. However, the limited convolution layer perception domain leads to the failure of global feature extraction, and consequently, the reconstructed image appears as a blurred overlay, or the recovered features do not correspond to the original image. Although expanding the convolution kernel size and deepening the convolution layers can help obtain the relevant global features, the computational complexity of the network increases exponentially. In recent years, attention mechanisms [41] have been widely used in natural language processing. In particular, the SAM module focuses selectively on significant location features to reduce network capacity and improve network computation performance. To obtain more global information, we designed a SAM module in the generator network (Figure 4).
The SAM module converts the previous feature nodes x R C × H × W into three mapping sub-layers: a pixel feature extraction space f ( x ) = W f x , a global feature extraction space g ( x ) = W g x , and a convolution feature mapping layer h ( x ) = W h x . To improve memory efficiency, the channel number C was set to C / k   k = 8 . We first transposed the multiplication between f ( x ) and g ( x ) . Subsequently, the transformed softmax classification function was used to obtain the feature attention map as follows:
β j , i = exp ( s i j ) i = 1 N exp ( s i j ) ,
where β j , i indicates the significance of the synthetic j-th region model for the i-th position, s i j = f ( x i ) T g ( x j ) , producing a series of attention map layers o = ( o 1 , o 2 , , o j , , o N ) R C × H × W , where one o j layer is shown as:
o j = i = 1 N β j , i h ( x i )
Finally, the attention map and feature mapping layer h ( x ) were added together to merge the globally relevant features. The output of the layer is expressed as follows:
y i = γ o i + x i ,
To explore local spatial information, γ was initialized as 0. Overall, self-attention can capture accurate image features beyond the convolution operation, and spatial connections between reference pixels across distances can help obtain more reference information. Furthermore, it also considers the key features of a high-resolution image with low computational complexity.

2.2.2. Weight Normalization

Batch Normalization (BN) [42] is a technique widely used in GAN-based methods. It balances the bias for image feature covariates inside the deep network and prevents overfitting during training. Nevertheless, BN tends to destroy low-frequency spatial features and increase erroneous noise estimates, causing artifacts in the SR processing results of remote sensing images. To overcome this issue, Salimans et al. have actively advocated for Weight Normalization (WN) [39] to stabilize the generative network, especially in GANs, by decoupling training weight vectors from their directions. WN decreases the dependence between mini-batch data and training performance, retains the contrast of image feature information, and has lower computational complexity than BN in equal time. Therefore, WN is more suitable than BN for performing SR tasks.
For nonlinear processing in the neural network, as seen in Equation (7), the former feature node x was computed with convolutional filtering with parameters w and b . New feature maps and nodes y were obtained by the activation function. WN in the convolution layer can be calculated using Equation (8):
y = f w x + b ,
w = g v v ,
where v and w have the same dimensional feature vector, and v denotes the Euclidean norm of v , which represents the direction of w . Scalar g determines the length w. After WN, the neural network nodes can be calculated as follows:
y = f g v x + b

2.3. Discriminator of TE-SAGAN

The role of the discriminator is to recognize the input images and judge their authenticity. The discriminator for TE-SAGAN is a VGGNet-type network structure [43] (Figure 5). This network has proven that multiple small convolution kernels (3 × 3) can learn more complex features at a lower computational cost than large convolutional kernels (e.g., 5 × 5, 7 × 7, and 11 × 11 in AlexNet) [44]. In this study, the fixed convolutional kernel size of VGGNet was modified to alternative kernel sizes of 3 and 4 to enhance the image feature recognition. Furthermore, the WN layers that replace all BN layers in the discriminator network can further improve the training network speed and stability.
The restored 128 × 128 SR and HR images were fed into the discriminator network. The image feature transferred through 10 layers of alternate convolution with kernel sizes of 3 × 3 and 4 × 4, step sizes of 1 and 2, and two linear dimensionality reduction operations. The channel numbers of the alternate convolution layers were 64, 128, 256, and 512. The generated feature map result was then sent into the final sigmoid function, outputting the authenticity of the remote sensing images.

2.4. Loss Function

The loss function measures the difference between the SR images generated by the network model and the HR images used for training the networks. In this study, to obtain a better result with more texture information and consistent style by TE-SAGAN, the loss function L ( I S R , H R ) during training network combined L1 norm content loss, training adversarial network loss, perceptual loss using a high-frequency network layer from VGGNet19, and texture loss computed using the Gram matrix.

2.4.1. Content Loss

Minimization of the MAE can enhance the sensitivity of abnormal image pixels. Therefore, it was used to describe the consistency of the low-frequency data between the recovered SR image and the original HR image. The definition of content loss is:
L con = L M A E ( I L R , θ ) = 1 M i = 1 M G ( I L R i ; θ ) I H R i ,
where I H R i and I L R i jointly form the i-th pair of the training sample set, and M represents the number of training dataset pairs. G ( I L R i ; θ ) is a mapping function between LR and SR images.

2.4.2. Adversarial Loss

The training stability of GANs is a measure used to ensure the quality of a generated image. The generator network “pretends to learn well” to cheat the discriminator and then yields terrible results. In this experiment, we employed the relativistic average standard GAN (RasGAN) [45], an adversarial training strategy with a relativistic discriminator, to compare the average input real data relative to faked data with a more realistic probability. The adversarial TE-SAGAN loss is unlike the rule of “either true or false” in the standard GAN discriminator. Instead, it focuses on the prediction of ground truth images to improve the network robustness. The generator and discriminator loss functions have mutual symmetry, and the corresponding loss functions of TE-SAGAN are defined in Equations (11) and (12):
L D R a = E I H R [ log ( D R a ( I H R , I S R ) E S R [ log ( 1 D R a ( I S R , I H R ) ) ] ,
L D R a = E I H R [ log ( 1 D R a ( I H R , I S R ) E S R [ log ( D R a ( I S R , I H R ) ) ] ,
where I H R denotes the original ground truth sample, I S R = G ( I L R ) , which is the generated SR remote sensing image, and E S R [ * ] represents the mean value of the generated remote sensing image for computation in the mini-batch. The adversarial loss of the generator is expressed as follows:
L adv = L G R a = E I H R [ ( 1 log ( D R a ( I H R , I S R ) ) ] E I S R [ ( log ( D R a ( I S R , I H R ) ) ] ,

2.4.3. Perceptual Loss

Typically, the pixel-wise loss (MSE or MAE) is restricted to the outcomes of low-frequency smooth features. However, it has some disadvantages for sparse and weak image feature values, leading to a low supervision capability for high-frequency features. To improve the quality of the generated HR remote sensing images and enhance the robustness of the deep network, perceptual loss was calculated using high-frequency feature maps before the activation operation, which can better represent the consistency of image texture information. In the pretrained VGG19 network [43], we calculated the Euclidean metric of high-frequency features extracted before the ReLU activation layers between the reference ground truth and the reconstructed SR image, defining their differences as perceptual loss:
L percep = 1 W i j H i j x = 1 W i , j y = 1 H i , j φ i j G ( I L R ) x , y φ i j I H R x , y 2 ,
where W i j H i j represents the dimension of the feature map in the VGGNet, I L R represents the LR remote sensing image, I H R represents the HR remote sensing image, φ i j indicates the activation operation of the feature map extracted before the j-th convolution layer and the i-th maximum pooling layer within the VGGNet19.

2.4.4. Texture Loss

Texture is also known as low-contrast fine detail. For SR remote sensing images, restoring the texture information to approach that of HR images is the primary objective. Consideration of perceptual loss improved the quality of the reconstructed image. However, the fine textures were still obscured. In this study, based on perceptual loss, the correlation between various feature channels was used to calculate texture loss. Texture loss can consolidate and correct the details of remote sensing images. Texture loss in HR-SR images can be calculated as follows:
L texture = G r ϕ I S R G r ϕ I H R 2 2 ,
where ϕ ( I ) R n × m represents the extracted features from VGGNet19, and G r is the feature map matrix computed on the gram matrix with the property of G r ( F ) = F F T R n × n .

2.4.5. Total Loss Function

Combined with the losses described in the previous sections, the total loss function is as follows:
L G = η L cont + λ L adv + L percep + β L texture ,
where L c o n t = L 1 = | | I S R I H R | | 1 and L a d v denote adversarial loss. L p e r c e p denotes the perceptual loss mentioned in Section 2.4.3. L t e x t u r e denotes the texture loss that describes the feature details. Accordingly, η , λ , and β are the coefficients that balance different sub-loss terms.

3. Experiment and Analysis

3.1. Evaluation Indexes

The peak signal-to-noise ratio (PSNR), structural similarity (SSIM) [46], and Fréchet Inception Distance (FID) [47] scores were used as objective evaluation indices for the experimental results, whereas the visual effect provided a subjective judgment. The PSNR was used to measure the pixel difference between the compressed image or signal reconstruction image and the original image in dB. A higher PSNR score indicates higher quality and pixel fidelity of the reconstructed image. The PSNR was calculated as follows:
M S E = 1 m n i = 0 m 1 j = 0 n 1 I ( i , j ) K ( i , j ) 2 ,
P S N R = 20 log 10 M A X I M S E ,
where M A X I represents the maximum pixel value ( M A X I is 255 for RGB images), and MSE represents the mean square error for images I and K with a size of m × n.
SSIM is a measure of luminance, contrast, and structure between two images in noise interference distortion. SSIM was calculated as follows:
S S I M x , y = 2 μ x μ y + C 1 2 σ x y + C 2 μ x 2 + μ y 2 + C 1 σ x 2 + σ y 2 + C 2 ,
where μ x and μ y represent the pixel mean of images x and y, respectively; σ x y represents the covariance of images x and y; naturally, σ x and σ y , respectively, represent the corresponding variance in images x and y. C 1 = ( K 1 L ) 2 and C 2 = ( K 2 L ) 2 are constants that avoid dividing by zero. Furthermore, L is the range of image pixel fluctuation. K 1 and K 2 are set orderly to 0.01 and 0.03 by default. SSIM consists of the following three components:
Luminance: l ( x , y ) = 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 ;
Contrast: c ( x , y ) = 2 σ x σ y + C 2 σ x 2 + σ y 2 + C 2 ;
Structure: s ( x , y ) = σ x y + C 3 σ x σ y + C 3 , with C 3 = C 2 / 2 .
In addition, the FID score, a proven systematic and convincing metric, was employed in this study to evaluate the accuracy of the designed model. FID computes the Fréchet distance (also known as Wasserstein-2 distance) between two Gaussian distributions (synthetic and real images) in the feature space [48]. FID constructs Gaussian model parameters around the matrix of two variables: the mean and covariance values ( m , C ) , which refer to the mean and covariance in HR ( m r , C r ) and generated images ( m g , C g ) . The FID score was determined as follows:
FID = d 2 ( ( m r , C r ) , ( m g , C g ) ) = m r m g 2 2 +   Tr ( C r + C g 2 ( C r C g ) ) ,
where Tr indicates the sum of the image features on the main diagonal of the square matrix. Accordingly, a lower FID value indicates that the produced image is closer to the HR sample.

3.2. Experiment Preparation

3.2.1. Data Set

The study area is located in the port city of Portsmouth, eastern Virginia, USA, and comprises water systems, houses, roads, and other landscape features. We built a dataset with 60,000 HR satellite image patches of 128 × 128 pixels without overlapping areas. The final 10,000 images were available for testing the trained model (called the Test-10000 set). Because the designed model’s aim is to obtain the single images’ super-resolution, the 4× super-resolution image processing is applied in this research, which is widely recognized in the current deep learning remote sensing image super-resolution tasks [49,50,51]. We used bicubic down-sampling with a common factor (4×) to obtain corresponding LR image patches and to establish a training dataset pair with HR image patches. In addition, given that super-resolution tasks primarily involve scaling of spatial resolution, the training dataset was only enhanced by random horizontal left and right flips, as well as 90°, 180°, and 270° rotations. Additionally, a well-known and publicly available dataset, the UC Merced Land Use dataset [52], was used in the test experiments. This dataset consisted of 256 × 256 pixels original HR image patches with 21 categories of satellite image scenes, each containing 100 patches of images. We selected the harbor, runway, airplane, and building scenes to validate the performance of the proposed method. Corresponding LR image patches of 64 × 64 pixels were obtained through the same down-sampling method.

3.2.2. Experimental Settings

All SR processing algorithms were implemented based on the TensorFlow framework in an Ubuntu 20.04.0 LTS system equipped with an NVIDIA GeForce RTX 2080 Ti GPU, CUDA10.1, and CUDNN7.6.0. As in a previous study [35], the training process was divided into two stages: pretraining and training. First, the proposed networks were trained with the L1 loss to consolidate the low-frequency image content and enhance the generator simulation capability. After pretraining, formal adversarial training was initialized with the parameters obtained from the final pretrained model. The balancing loss term coefficients in Equation (16) were η = 10 2 , λ = 5 × 10 4 , and β = 10 6 . The batch size of the imported training data was set to 16 in the pretraining phase and 4 in the adversarial training phase. The initial learning rate was set at 2 × 10 4 , and half-decayed every 2 × 10 5 during the pretraining phase and every 5 × 10 4 during the training phase. Adam was used for the overall training optimization [53] in the training stage with default hyperparameters ( β 1 = 0.9, β 2 = 0.999).

3.3. Results and Analysis

3.3.1. Quantitative Evaluation

First, by comparing the results obtained using enhanced super-resolution GAN (ESRGAN) [35] and TE-SAGAN in the pretraining phase, we determined the generator network stabilization performance of the two networks. A visual example of PSNR values in the entire pretraining stage for the two methods is displayed in Figure 6. The results revealed that, compared with the baseline, the pretraining PSNR value for TE-SAGAN was slightly lower, and the network stability was considerably stronger without heavy fluctuations owing to the well-trained SAM module and WN layers.
To validate the pretrained model in detail, further experiments were conducted using remote sensing images of different scenes in the UC Merced Land Use dataset. As shown in Table 1, our method yielded better results than ESRGAN. Simultaneously, as depicted in Figure 7, the current loss function only described the low-frequency content of the image, resulting in smooth outcomes that were still not entirely convincing. We further compared several reconstructed tree shadow textures in detail and concluded that the results of the proposed model were closer to the ground truth. However, both methods produced smooth and fuzzy images in general. For alignment with human visual perception, further training using an adversarial component based on the pretrained model is necessary.
To complement prior studies, bicubic interpolation and four representative deep-learning-based SR technologies, namely, ESPCN [30], EDSR [54], SRGAN [34], ESRGAN [35], and RFDNet [55], were used as comparison methods. The quantitative comparison results are presented in Table 2. However, the mean PSNR values of our method were only 0.050 and 0.398 dB higher than the second-highest RFDNet algorithm for Test-10000 and runway, respectively. Moreover, the remaining mean PSNR values of our model were lower than RFDNet by 0.990, 0.234, and 0.205 dB for the harbor, airplane, and buildings sub-datasets, respectively, and all average SSIM scores of our model were lower than highest RFDNet by 0.017, 0.042, 0.008, 0.019, and 0.026. Instead, in the FID score term, where the lower score implies that generated images are closer to real reference images, most of the FID scores in our network were significantly lower than those in the second-lowest network, ESRGAN, by 8.816, 21.749, 7.139, and 10.580 for Test-10000, runway, airplane, and buildings, respectively, expect for the harbor sub-dataset, which was higher than RFDNet by 6.28. These findings indicate that compared with the other methods, TE-SAGAN focuses on sharpening the super resolution image while keeping a relatively high level of PSNR and SSIM image fidelity evaluations. Our method obtains a better objective rating for the four benchmark datasets. The program runtime is a non-negligible metric for evaluating network performance. ESPCN takes the shortest runtime with fast training using sub-pixel convolution, and the recent RFDNet belongs to an improved lightweight distillation network with a relatively low training runtime. Among the models with runtime lasting more than one day (1440 min), TE-SAGAN takes the shortest training time, which shortens around a quarter of the runtime than ESRGAN. The weight normalization (WN) layer introduced in this paper ensures the network stability in inputting mini-batch data to accelerate training time. Training EDSR takes the longest time. These findings indicate that TE-SAGAN shows improvement in network runtime.
To systematically elucidate how the introduced WN and SAM affect the network performance, several TE-SAGAN ablation experiments were performed using the following networks: baseline (ESRGAN), baseline combined with the WN layer (Baseline + WN), baseline combined with the WN layer and SAM (Baseline + WN + SAM), and the final combination of SAM, WN layer, and texture loss term (Baseline + WN + SAM + Texture). Based on the results presented in Figure 8, the following conclusions were drawn. First, the addition of the WN and SAM modules to the baseline network improved the PSNR (Figure 8a) and SSIM (Figure 8b) values for several sub-datasets in the UC Merced Land Use dataset. This implies that the added modules are effective in improving the quality and fidelity of the reconstructed image. Second, the introduced texture loss focuses more on restoring texture details (FID score in Figure 8c) than on enhancing PNSR and SSIM values. Furthermore, although our network parameter was slightly larger than that of the baseline model during the experiment, our network showed higher training efficiency. In summary, the ability of WN and SAM modules to rapidly and accurately extract image features allows TE-SAGAN to reconstruct images with high fidelity, and texture loss further adds details consistent with human observation to the reconstructed image. We further believe that the added modules are associated with improved network fitness.

3.3.2. Qualitative Evaluation

Faithful SR images has always been the focus of many image-processing tasks. As mentioned above, the PSNR and SSIM metrics obtained with the networks were different from human observations. Therefore, to further improve the texture and edge of reconstructed remote sensing images, qualitative visual comparisons were performed (Figure 9 and Figure 10). The overall quality of SR remote sensing images obtained using our method was significantly higher than those obtained via other algorithms. Figure 9 shows the processing results of our method for the test dataset (Test-10000), including the villa surroundings, car windows in the parking lot, and shade trees beside the road. In detail, bicubic interpolation [14] filled pixel information with a fixed computational paradigm, producing a blurry image. SRGAN [34] and ESRGAN [35] yielded more visually pleasant processing images than ESPCN [30] and EDSR [54] owing to their excellent depth feature extraction module and perceptual loss. However, the textures recovered by SRGAN and ESRGAN were inadequate and erroneous (e.g., car windows and tree shadows) because they lack expression and constraints on detailed and realistic textures. In particular, RFDNet is biased with respect to objective PSNR and SSIM showing higher scores, but generating super-resolution images with unclear textures. Our method was equipped with the WN layer and the Gram matrix for texture loss calculation, which significantly increased the calculation of high-frequency information correlation between SR and HR images and limited the pixel and texture deviation. Thus, our method yielded finer-grained and more realistic images.
To examine the recovery status of image outlines, experiments were conducted using four representative remote sensing images from the UC Merced Land Use dataset, including street corners, rooftops, harbor ships, traffic lines, and other scenes, and the results are presented in Figure 10. SRGAN and ESRGAN produced more realistic edges than bicubic, ESPCN, EDSR, and RFDNet. However, the textures recovered by SRGAN and ESRGAN still contained unreal artifacts, as they rarely consider the spatial information of remote sensing images. In particular, RFDNet obtained high scores in objective PSNR and SSIM but generated super-resolution images with unclear outlines. In contrast, the SAM module in our proposed TE-SAGAN enhanced the connection between distant pixels, resulting in clearer, sharper edges in the reconstructed remote sensing images. Interestingly, even though several of the original HR images were slightly blurry, our method produced images with clearer and more natural borders than the original.

4. Conclusions

SR images have been widely applied for extracting information from remote sensing images. In this study, we elucidated why some algorithms lose control over edges and internal textures in the reconstructed remote sensing images even after smoothing. We proposed an RRDB-based texture-enhancement network, TE-SAGAN, which integrates the SAM and WN into the GAN framework. The SAM module effectively combines wide-area and long-range image information training. The WN layer stabilizes the reconstructed image quality using mini-batch data, thereby reducing the contrast gaps and artifacts in the generated images. In addition, the Gram matrix imposes a texture loss constraint to refine the reconstructed texture. The results for PSNR, SSIM, and FID scores of TE-SAGAN during training using our own and public benchmark remote sensing image datasets indicate improved performance. This study also suggests that our proposed method yields clearer image edges and more realistic internal details compared to existing methods. Of note, the main limitation of our current study is that we focused on rebuilding texture details and borders by increasing the model parameters. Further exploratory research is needed to reduce the model size while achieving equivalent or even superior results.

Author Contributions

Y.X. and L.T. proposed the architecture design. Y.X. and A.H. performed the experiments and analyzed the data. Y.X. and W.L. wrote the paper. X.X. and Z.X. proposed the figure in this paper. Z.X. and A.H. revised the paper and provided valuable advice for the experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This study is funded by the National Natural Science Foundation of China (42001340) and the Open Fund of Key Laboratory of Urban Land Resources Monitoring and Simulation, Ministry of Natural Resources (KF-2020-05-068).

Data Availability Statement

This study did not report any data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xu, Y.; Wu, L.; Xie, Z.; Chen, Z. Building extraction in very high resolution remote sensing imagery using deep learning and guided filters. Remote Sens. 2018, 10, 144. [Google Scholar] [CrossRef] [Green Version]
  2. Xu, Y.; Xie, Z.; Feng, Y.; Chen, Z. Road extraction from high-resolution remote sensing imagery using deep learning. Remote Sens. 2018, 10, 1461. [Google Scholar] [CrossRef] [Green Version]
  3. Guo, M.; Liu, H.; Xu, Y.; Huang, Y. Building extraction based on U-Net with an attention block and multiple losses. Remote Sens. 2020, 12, 1400. [Google Scholar] [CrossRef]
  4. Wei, S.; Ji, S. Graph convolutional networks for the automated production of building vector maps from aerial images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5602411. [Google Scholar] [CrossRef]
  5. Xia, L.; Zhang, X.; Zhang, J.; Wu, W.; Gao, X. Refined extraction of buildings with the semantic edge-assisted approach from very high-resolution remotely sensed imagery. Int. J. Remote Sens. 2020, 41, 8352–8365. [Google Scholar] [CrossRef]
  6. Tom, B.C.; Katsaggelos, A.K. Reconstruction of a High-Resolution Image by Simultaneous Registration, Restoration, and Interpolation of Low-Resolution Images. Proc. Int. Conf. Image Process. 1995, 2, 539–542. [Google Scholar]
  7. Galbraith, A.E.; Theiler, J.; Thome, K.J.; Ziolkowski, R.W. Resolution Enhancement of Multilook Imagery for the Multispectral Thermal Imager. IEEE Trans. Geosci. Remote Sens. 2005, 43, 1964–1977. [Google Scholar] [CrossRef]
  8. He, Y.; Yap, K.-H.; Chen, L.; Chau, L.-P. A Soft MAP Framework for Blind Super-Resolution Image Reconstruction. Image Vis. Comput. 2009, 27, 364–373. [Google Scholar] [CrossRef]
  9. Haut, J.M.; Fernández-Beltran, R.; Paoletti, M.E.; Plaza, J.; Plaza, A.J.; Pla, F. A New Deep Generative Network for Unsupervised Remote Sensing Single-Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6792–6810. [Google Scholar] [CrossRef]
  10. Lei, S.; Shi, Z.; Zou, Z. Coupled Adversarial Training for Remote Sensing Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2020, 58, 3633–3643. [Google Scholar] [CrossRef]
  11. Tsai, R.Y.; Huang, T.S. Multiframe Image Restoration and Registration. Adv. Comput. Vis. Image Process. 1984, 1, 317–339. [Google Scholar]
  12. Cover, T.; Hart, P. Nearest Neighbor Pattern Classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  13. Duchon, C.E. Lanczos Filtering in One and Two Dimensions. J. Appl. Meteorol. Climatol. 1979, 18, 1016–1022. [Google Scholar] [CrossRef]
  14. Carlson, R.E.; Hall, C.A. Error Bounds for Bicubic Spline Interpolation. J. Approx. Theory 1973, 7, 41–47. [Google Scholar] [CrossRef] [Green Version]
  15. Miles, N. Method of Recovering Tomographic Signal Elements in a Projection Profile or Image by Solving Linear Equations. JUSTIA Patents No. 5323007, 21 June 1994. [Google Scholar]
  16. Stark, H.; Olsen, E.T. Projection-Based Image Restoration. J. Opt. Soc. Am. A-Opt. Image Sci. Vis. 1992, 9, 1914–1919. [Google Scholar] [CrossRef]
  17. Stark, H.; Oskoui, P. High-Resolution Image Recovery from Image-Plane Arrays, Using Convex Projections. J. Opt. Soc. Am. A Opt. Image Sci. 1989, 6, 1715–1726. [Google Scholar] [CrossRef]
  18. Irani, M.; Peleg, S.; Chang, H.; Yeung, D.; Xiong, Y. Super Resolution from Image Sequences Super-Resolution through Neighbor Embedding. CVPR 1990, 2, 115–120. [Google Scholar]
  19. Unser, M.A.; Aldroubi, A.; Eden, M. Fast B-Spline Transforms for Continuous Image Representation and Interpolation. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 277–285. [Google Scholar] [CrossRef]
  20. Unser, M.A.; Aldroubi, A.; Eden, M. B-Spline Signal Processing: Part I—Theory. IEEE Trans. Signal Process. 1993, 41, 821–833. [Google Scholar] [CrossRef]
  21. Unser, M.A.; Aldroubi, A.; Eden, M. B-Spline Signal Processing: Part II-Efficient Design and Applications. IEEE Trans. Signal Process. 1993, 41, 834–848. [Google Scholar] [CrossRef]
  22. Xu, Y.; Jin, S.; Chen, Z.; Xie, X.; Hu, S.; Xie, Z. Application of a graph convolutional network with visual and semantic features to classify urban scenes. Int. J. Geogr. Inf. Sci. 2022, 1–26. [Google Scholar] [CrossRef]
  23. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the ECCV 2014, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  24. Denton, E.L.; Chintala, S.; Szlam, A.D.; Fergus, R. Deep Generative Image Models Using a Laplacian Pyramid of Adversarial Networks. NIPS 2015, 28, 1486–1494. [Google Scholar]
  25. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 694–711. [Google Scholar]
  26. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-Recursive Convolutional Network for Image Super-Resolution. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  27. Li, J.; Fang, F.; Mei, K.; Zhang, G. Multi-Scale Residual Network for Image Super-Resolution. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 527–542. [Google Scholar]
  28. Han, W.; Chang, S.; Liu, D.; Yu, M.; Witbrock, M.; Huang, T.S. Image Super-Resolution via Dual-State Recurrent Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1654–1663. [Google Scholar]
  29. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  31. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
  32. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A Persistent Memory Network for Image Restoration. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  33. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image Super-Resolution Using Dense Skip Connections. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4809–4817. [Google Scholar]
  34. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; Shi, W. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 June 2017; pp. 105–114. [Google Scholar]
  35. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Loy, C.C.; Qiao, Y.; Tang, X. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the ECCV Workshops 2018, Munich, Germany, 8–14 September 2018. [Google Scholar]
  36. Rakotonirina, N.C.; Rasoanaivo, A. ESRGAN+: Further Improving Enhanced Super-Resolution Generative Adversarial Network. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3637–3641. [Google Scholar]
  37. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 10 October 2021; pp. 1905–1914. [Google Scholar]
  38. Jo, Y.; Yang, S.; Kim, S.J. Investigating Loss Functions for Extreme Super-Resolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1705–1712. [Google Scholar]
  39. Salimans, T.; Kingma, D.P. Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2016; pp. 901–909. [Google Scholar]
  40. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Bangkok, Thailand, 18–22 November 2020; MIT Press: Cambridge, MA, USA, 2020; Volume 2, pp. 2672–2680. [Google Scholar]
  41. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6000–6010. [Google Scholar]
  42. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on International Conference on Machine Learning, Lille, France, 7–9 July 2015; JMLR.org: Lille, France, 2015; Volume 37, pp. 448–456. [Google Scholar]
  43. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  44. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  45. Jolicoeur-Martineau, A. The Relativistic Discriminator: A Key Element Missing from Standard GAN. arXiv 2018, arXiv:1807.00734. [Google Scholar]
  46. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  47. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 6629–6640. [Google Scholar]
  48. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  49. Ma, W.; Pan, Z.; Guo, J.; Lei, B. Super-resolution of remote sensing images based on transferred generative adversarial network. In Proceedings of the IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 1148–1151. [Google Scholar]
  50. Zhang, Z.; Tian, Y.; Li, J.; Xu, Y. Unsupervised Remote Sensing Image Super-Resolution Guided by Visible Images. Remote Sens. 2022, 14, 1513. [Google Scholar] [CrossRef]
  51. Guo, M.; Zhang, Z.; Liu, H.; Huang, Y. NDSRGAN: A Novel Dense Generative Adversarial Network for Real Aerial Imagery Super-Resolution Reconstruction. Remote Sens. 2022, 14, 1574. [Google Scholar] [CrossRef]
  52. Yang, Y.; Newsam, S. Bag-of-Visual-Words and Spatial Extensions for Land-Use Classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, New York, NY, USA, 2–5 November 2010; Association for Computing Machinery: New York, NY, USA, 2010; pp. 270–279. [Google Scholar]
  53. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. Int. Conf. Learn. Represent. 2014. [Google Scholar] [CrossRef]
  54. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  55. Liu, J.; Tang, J.; Wu, G. Residual Feature Distillation Network for Lightweight Image Super-Resolution. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Bartoli, A., Fusiello, A., Eds.; Springer International Publishing: Glasgow, UK, 2020; Volume 12537, pp. 41–55. [Google Scholar]
Figure 1. Schematic diagram of the super-resolution image processing in TE-SAGAN.
Figure 1. Schematic diagram of the super-resolution image processing in TE-SAGAN.
Remotesensing 14 02425 g001
Figure 2. Structure of the generator network of TE-SAGAN.
Figure 2. Structure of the generator network of TE-SAGAN.
Remotesensing 14 02425 g002
Figure 3. Structure diagram of RRDB (convolution kernel size k = 3, stride s = 1, scaling factor α = 0 . 2 ).
Figure 3. Structure diagram of RRDB (convolution kernel size k = 3, stride s = 1, scaling factor α = 0 . 2 ).
Remotesensing 14 02425 g003
Figure 4. The self-attention mechanism module structure.
Figure 4. The self-attention mechanism module structure.
Remotesensing 14 02425 g004
Figure 5. Schemes follow the same formatting. Structure diagram of the discriminator network, k denotes the convolutional kernel size, n denotes the number of convolutional kernel channels, and s denotes the convolutional stride.
Figure 5. Schemes follow the same formatting. Structure diagram of the discriminator network, k denotes the convolutional kernel size, n denotes the number of convolutional kernel channels, and s denotes the convolutional stride.
Remotesensing 14 02425 g005
Figure 6. PSNR values in the pretraining networks. Baseline: pre-ESRGAN, the pretrain for ESRGAN; ours: pre-TE-SAGAN.
Figure 6. PSNR values in the pretraining networks. Baseline: pre-ESRGAN, the pretrain for ESRGAN; ours: pre-TE-SAGAN.
Remotesensing 14 02425 g006
Figure 7. Results of SR processing using the Test-10000 dataset with a scale factor of 4 in the pretraining stage.
Figure 7. Results of SR processing using the Test-10000 dataset with a scale factor of 4 in the pretraining stage.
Remotesensing 14 02425 g007
Figure 8. Ablation results of PSNR (a) , SSIM (b) , and FID (c) values using the UC Merced Land Use dataset. Baseline, ESRGAN; Baseline + WN, baseline combined with the WN layer; Baseline + WN + SAM, baseline combined with the WN and SAM modules; Baseline + WN + SAM + Texture, final combination of the SAM module, WN layer, and texture loss term.
Figure 8. Ablation results of PSNR (a) , SSIM (b) , and FID (c) values using the UC Merced Land Use dataset. Baseline, ESRGAN; Baseline + WN, baseline combined with the WN layer; Baseline + WN + SAM, baseline combined with the WN and SAM modules; Baseline + WN + SAM + Texture, final combination of the SAM module, WN layer, and texture loss term.
Remotesensing 14 02425 g008
Figure 9. Image processing results of the different algorithms using the test dataset, including the details of roof, cars, and tree shadows.
Figure 9. Image processing results of the different algorithms using the test dataset, including the details of roof, cars, and tree shadows.
Remotesensing 14 02425 g009
Figure 10. Comparative results of super-resolution image using the UC Merced Land Use dataset.
Figure 10. Comparative results of super-resolution image using the UC Merced Land Use dataset.
Remotesensing 14 02425 g010
Table 1. Comparison of pre-ESRGAN and Pre-TE-SAGAN using part of the UC Merced Land Use dataset.
Table 1. Comparison of pre-ESRGAN and Pre-TE-SAGAN using part of the UC Merced Land Use dataset.
DatasetIndexPre-ESRGANPre-TE-SAGAN
HarborSSIM0.8000.798
PSNR23.25023.154
RunwaySSIM0.8190.821
PSNR30.57430.724
AirplaneSSIM0.8190.821
PSNR28.74828.800
BuildingsSSIM0.8050.806
PSNR26.12426.138
Highest measures (PSNR [dB], SSIM) in bold [up-sampling scale 4×].
Table 2. Comparison of Bicubic, ESPCN, EDSR, SRGAN, ESRGAN, RFDNet and TE-SAGAN on the Test-10000 set and part of UC Merced Land Use.
Table 2. Comparison of Bicubic, ESPCN, EDSR, SRGAN, ESRGAN, RFDNet and TE-SAGAN on the Test-10000 set and part of UC Merced Land Use.
DatasetMetricBicubicESPCNEDSRSRGANESRGANRFDNetTE-SAGAN
Test-10000SSIM0.5580.5000.5640.5390.5750.6000.583
PSNR22.85422.14122.85522.69523.50023.65023.700
FID136.312148.19995.22655.25232.58781.86623.771
HarborSSIM0.7160.6350.6680.6180.6590.7490.707
PSNR20.31419.33919.30218.81419.76521.76020.770
FID147.225169.257205.605227.557135.349115.247121.527
RunwaySSIM0.7040.6680.6780.7080.6980.7560.748
PSNR26.01025.39326.86326.79426.95928.06928.467
FID146.718191.391132.028116.549105.031106.09583.282
AirplaneSSIM0.7130.6620.7090.6770.6690.7520.733
PSNR25.42123.86124.56123.99224.84926.5326.296
FID179.435201.591158.942143.52189.532119.77382.393
BuildingsSSIM0.6990.6250.6700.6430.6560.7460.720
PSNR23.22621.69522.13721.78722.55924.2124.005
FID127.26140.786149.487122.7881.284118.87070.704
Runtime (min)/107.83225.62292.03072.9251.82288.7
For PSNR [dB] and SSIM , the highest values are in bold, whereas the second-highest values are underlined. On the contrary, for FID , the lowest values are in bold, whereas the second-lowest values are underlined.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xu, Y.; Luo, W.; Hu, A.; Xie, Z.; Xie, X.; Tao, L. TE-SAGAN: An Improved Generative Adversarial Network for Remote Sensing Super-Resolution Images. Remote Sens. 2022, 14, 2425. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14102425

AMA Style

Xu Y, Luo W, Hu A, Xie Z, Xie X, Tao L. TE-SAGAN: An Improved Generative Adversarial Network for Remote Sensing Super-Resolution Images. Remote Sensing. 2022; 14(10):2425. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14102425

Chicago/Turabian Style

Xu, Yongyang, Wei Luo, Anna Hu, Zhong Xie, Xuejing Xie, and Liufeng Tao. 2022. "TE-SAGAN: An Improved Generative Adversarial Network for Remote Sensing Super-Resolution Images" Remote Sensing 14, no. 10: 2425. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14102425

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop