Next Article in Journal
Deep Learning for Precipitation Estimation from Satellite and Rain Gauges Measurements
Next Article in Special Issue
Road Extraction of High-Resolution Remote Sensing Images Derived from DenseUNet
Previous Article in Journal
Using Predictive and Differential Methods with K2-Raster Compact Data Structure for Hyperspectral Image Lossless Compression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Multi-Scale Recurrent Network for Synthetic Aperture Radar Images Despeckling

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(21), 2462; https://0-doi-org.brum.beds.ac.uk/10.3390/rs11212462
Submission received: 23 September 2019 / Revised: 16 October 2019 / Accepted: 19 October 2019 / Published: 23 October 2019

Abstract

:
For the existence of speckles, many standard optical image processing methods, such as classification, segmentation, and registration, are restricted to synthetic aperture radar (SAR) images. In this work, an end-to-end deep multi-scale recurrent network (MSR-net) for SAR image despeckling is proposed. The multi-scale recurrent and weights sharing strategies are introduced to increase network capacity without multiplying the number of weights parameters. A convolutional long short-term memory (convLSTM) unit is embedded to capture useful information and helps with despeckling across scales. Meanwhile, the sub-pixel unit is utilized to improve the network efficiency. Besides, two criteria, edge feature keep ratio (EFKR) and feature point keep ratio (FPKR), are proposed to evaluate the performance of despeckling capacity for SAR, which can assess the retention ability of the despeckling algorithm to edge and feature information more effectively. Experimental results show that our proposed network can remove speckle noise while preserving the edge and texture information of images with low computational costs, especially in the low signal noise ratio scenarios. The peak signal to noise ratio (PSNR) of MSR-net can outperform traditional despeckling methods SAR-BM3D (Block-Matching and 3D filtering) by more than 2 dB for the simulated image. Furthermore, the adaptability of optical image processing methods to real SAR images can be enhanced after despeckling.

Graphical Abstract

1. Introduction

Synthetic aperture radar (SAR), owing to its all-weather and all-time condition operation, has been widely applied to microwave remote sensing areas, such as topographic mapping, military target reconnaissance, and natural disaster monitoring [1,2]. SAR imaging achieves high range resolution by exploiting pulse compression technique and high azimuth resolution by using radar platform to form a virtual antenna synthetic aperture along track [3,4]. However, speckle noise exists in the imaging results due to the coherent imaging mechanism of SAR, which leads to images quality and readability reduction. Meanwhile, the existence of speckles limits the effectiveness of the application of common optical image processing methods to SAR images [5]. It thus restricts the SAR images to further understanding and interpretation, increasing the difficulty of extracting roads, farmlands, and buildings in the image and the complexity of spatial feature extraction in image registration, and reducing the accuracy of detection and classification of the objects such as vehicles and ships [6]. Speckle suppression is, therefore, an important task in SAR image post-processing.
To improve the quality of SAR images, there have been various speckle suppression methods proposed, including multi-look processing technologies during imaging and image filtering methods after imaging [1,7]. Multi-look processing divides the whole effective synthetic aperture length into multiple segments. The incoherent sub-views are then superimposed to obtain the high signal-to-noise ratio (SNR) images [8]. However, multi-look processing reduces the utilization of Doppler bandwidth, resulting in a decrease of the spatial resolution of the imaging results, which cannot meet the requirements of high resolution [9].
The filtering methods are mainly divided into three categories: the spatial filtering method, the transform domain filtering method, and the non-local mean filtering method. Median filtering and mean filtering are the earliest spatial filtering methods of traditional digital image processing. Although these two methods can suppress speckles to a certain extent, it leads to image blurring and objects edge information loss. Afterward, Lee filter [10], Frost filter [11], and Kuan filter [12] are designed for speckles suppression of SAR images. Based on the coherent speckle multiplier model, Lee filter selects a fixed local window in the image, assuming that the prior mean and variance can be calculated by the local region [10]. This method has a small amount of computation, but the selection of local window size has a great influence on the result, and the details and edge information of the image may be lost [13]. Frost filter assumes that the SAR image is a stationary random process and coherent spot noise is multiplying noise, and uses the least mean square error criterion to estimate the real image [11]. For the reason that the actual SAR image does not fully meet the hypothesis, the SAR image processed by this method will have blurred edges in areas with rich details. Kuan filters to apply sliding windows to estimate the local statistical properties of the image and then replaces the global characteristics of the image with these local statistical properties [12].
The representatives of transform domain filtering methods are threshold filtering method and multi-resolution analysis method based on wavelet transform. Donoho et al. first proposed a hard threshold and soft threshold denoising method based on wavelet transform [14]. After that, Bruce and Gao et al. proposed semi-soft threshold function methods to improve the hard threshold and soft threshold denoising methods [15,16]. This method solved the problems of discontinuity of hard threshold function and constant deviation of reconstructed signal with soft threshold function. He et al. later proposed a wavelet Markov model [17], which achieved a significant result in SAR images denoising. However, the wavelet transform cannot deal with two- and higher-dimensional images well. Because in the case of high-dimensional wavelet basis, one-dimensional wavelet basis cannot obtain the optimal representation of the two-dimensional function. In recent years, the appearance of multi-scale geometric analysis has made up this defect. Also, multi-scale geometric analysis tools are abundant, such as Ridgelets transformation, Curvelets transformation, Brushlets transformation, and Contourlets transformation [18,19,20]. These transform domain filtering methods have good coherent spots suppression capability, which preserved image details and edge information while speckles are removed. However, the processing of these algorithms is based on local characteristics, which is complex with a large amount of computation and easily producing pseudo-Gibbs stripes.
The non-Local Means (NLM) filtering method [21] proposed by Buades et al. repeatedly searched the whole image with similar image blocks and used similar texture regions instead of noise regions to achieve denoising. [The authors of [22,23] applied NLM to SAR image denoising, which can effectively eliminate the speckles. Kervrann et al. [24] further improved NLM by proposing a new adaptive algorithm that modified the similarity measurement parameters of NLM. Dabov et al. [22] proposed a Block-Matching and 3D filtering (BM3D) algorithm which applied the local linear minimum mean variance (MMSE) criterion and wavelet transform, and combined the non-local mean idea with the transform domain filtering method. It is one of the best methods for denoising at present. However, this algorithm needs a large number of search operations at the cost of a large amount of computation and low efficiency.
In recent years, deep convolutional neural network (CNN) has developed rapidly, which provides a new idea for SAR image despeckling. Wang et al. constructed an image despeckling convolutional neural network (ID-CNN) in [25]. ID-CNN can directly estimate the speckle distribution and eliminate the estimated speckles from the image to obtain a clean image. Different from the ID-CNN, Yue et al. in [6], combining the statistical model with CNN, proposed a framework that does not require reference images and could work in an unsupervised way when trained with real SAR images. Bai et al. [26] added fractional total variational loss to the loss function to remove the obvious noise while maintaining the texture details. [The authors of [27] proposed a CNN framework based on dilated convolutions called SAR-DRN. This network amplified the receptive field by dilated convolutions and further improved the network by exploiting the skip connections and a residual learning strategy. State-of-the-art results are achieved in both quantitative and visual assessments.
In this study, we design an end-to-end multi-scale recurrent network for SAR image despeckling. Unlike [9,25,26,27], which only utilized CNN to acquire speckle distribution characteristics and additional division operation or subtraction operation to remove speckle, we use the network to learn the distribution characteristics of speckle noise, meanwhile automatically implementing speckle suppression to output clean images. The proposed network is based on the encoder–decoder architecture. To improve the operation efficiency, in the decoder part, we use the subpixel unit to implement up-sample on the feature maps instead of the deconvolutional layer. Besides, this paper applies a multi-scale recurrent strategy, which inputs the resized images with different scales to the network, and different scale inputs share the same network weights parameters. So, the network performance can be improved without increasing the network parameters and the output which is friendly to the optical image processing algorithm can be obtained. Also, the convolutional LSTM unit is used to implement information transmit among each scale. Although our network is the same as the network based on noise output, i.e., a fully convolutional network, our MSR-net contains the pooling layer that can reduce the dimension of the network and further reduce the amount of computation to a great extent. Lastly, we propose two evaluation criteria based on image processing methods.
The paper consists of 6 parts. In Section 2, we analyze the speckles of SAR images and briefly introduce CNN, and convolutional LSTM. After providing the framework of our proposed MSR-net in Section 3, the result and discussion of the experiment are shown in Section 4 and Section 5. The last section will summarize this paper.

2. Review of Speckle Model and Neural Network

2.1. Speckle Model of Sar Images

Multiplicative model is usually used to describe speckle noise [28] and the formula is defined as:
I = p s · n ,
where I is image intensity, p s is a constant which denotes the average scattering coefficient of objects or ground, and n denotes the speckle which is independent with p s statistically.
For the homogeneous SAR image, the single-look intensity I obeys negative exponential distribution [29] and its probability distribution function (PDF) is defined as:
p ( I ) = 1 p s exp I p s .
The multi-look processing methods are usually used to improve the quality of SAR images by diminishing the speckle noise. If the Doppler bandwidth is divided into L sub-bandwidths during imaging, and I i is the single-look intensity image corresponding to each sub-bandwidth, the result of multi-look processing is:
I = 1 L i = 1 L I i ,
where L is the number of looks. If I i obeys the exponential distribution in Equation (2), then after multi-look averaging, the L-look intensity image follows the Gamma distribution [1], and the PDF is:
p I = 1 p s Γ ( L ) L I p s L 1 exp L I p s , L 1 ,
where Γ ( L ) denotes the Gamma function. The PDF of L-look speckle n can be obtained by applying the product model on Equations (1) and (4),
p ( n ) = L L n L 1 exp ( L n ) Γ ( L ) , L 1 .

2.2. Convolutional Long Short-Term Memory

Convolutional neural networks (CNNs) have powerful capabilities of extracting spatial features and can automatically extract universal features through back-propagation algorithms driven by dataset [30,31], however they cannot be used to process sequence signals directly for the reason that the input is independent with each other, and the information flows strictly in one direction from layer to layer.To solve this problem, we introduce convolutional long short-term memory (ConvLSTM) [32] to the network, the inner structure of ConvLSTM is shown in Figure 1. As a special kind of RNN, long short-term memory network (LSTM) has internal hidden memory which allows the model to store information about its past computations and capable of learning long-term dependency [33]. Different than standard LSTM, all of the features variables of ConvLSTM including the input X t , cell state C t , the output of the forget gate F t , input gate i t , and output gate O t are three-dimensional tensors, the latter two of which are spatial dimensions width and height. The key equations of ConvLSTM are defined as:
F t = σ ( W f [ H t 1 , X t ] + b f ) , i t = σ ( W i [ H t 1 , X t ] + b i ) , C ˜ t = tanh ( W C [ H t 1 , X t ] + b c ) , C t = F t * C t 1 + i t * C ˜ t , O t = σ ( W o [ H t 1 , F t ] + b o ) , H t = O t * tanh ( C t ) ,
where “*" and “ σ " denote the Hadamard product and the logistic sigmoid function, respectively. F t controls the abandoned state information of the last layer and i t is in charge of current state update, i.e., C ˜ t . W f , W i , W c , and W o represent weights of each neural unit with b f , b i , b c , and b o denoting the corresponding offsets.

3. Proposed Method

An end-to-end network MSR-net for SAR image despeckling is proposed in this paper. Rather than using additional division operation [25,26] or subtraction operation [9,27], our network can automatically perform despeckling and generate a clean image. In this section, we first introduce the multi-scale recurrent architecture and then describe specific details through a single-scale model.

3.1. Architecture

MSR-net is built based on cascaded subnetworks, and each subnetwork contains three parts: encoder, decoder, and ConvLSTM unit, as illustrated in Figure 2. Different levels of subnetworks correspond to different scales of inputs and outputs. The next scale speckled image and the output of current subnetwork are combined as the input of next-level subnetwork. In addition, an LSTM unit with single input and two-output is embedded between encoder and decoder. Specifically, one output is connected to the decoder, and the other output which represents the hidden state is connected to LSTM unit of the next subnetwork.
Different from the general cascaded network like [34], which uses three stages of independent subnetworks, all the state features flow across scales and share the same training parameters in MSR-net. Owing to the multi-scale recurrent and parameter share strategy, the number of parameters that need to be trained in MSR-net is only 1/3 of [34].
For the subnetwork, the output F i of encoder Net e n , which takes the speckled image and despeckled result up-sampled from the previous scale as input, can be defined as:
F i = Net e n I i n i , u p ( I o i + 1 ) ; Θ e n ,
where I i n i is the input image with speckle noise, Θ e n is the weights parameters of Net e n . i = 0 , 1 , 2 , . . . is the scale index. The larger i is, the lower the resolution is. i = 0 represents the original resolution and i = 1 indicates down sampling once. I o i + 1 is the output of the previous coarse scale. u p ( · ) is the operator that adapts features or images from the (i + 1)-th to the i-th scale, which is implemented by bilinear interpolation.
To exploit the information contained in feature maps of different scales, a convolutional LSTM module is embedded between the encoder and the decoder. The ConvLSTM can be defined as:
h i , g i = ConvLSTM u p ( h i + 1 ) , F i ; Θ L S T M ,
where Θ L S T M is the set of parameters in ConvLSTM, h i is the hidden state which is passed to the next scale, h i + 1 is the hidden state from the previous scales, and g i is the output of the current state (scale). Finally, we use Θ d e to denote the parameters of the decoder, and the output can be defined as:
I o i = Net d e g i ; Θ d e .

3.2. Single Scale Network

Details of the MSR-net are introduced by the single-scale model in this section. As shown in Figure 3, the single-scale model consists of two parts: encoder and decoder. The encoder includes three building blocks: convolutional layer, pooling layer, and Res block.
The convolution unit performs convolution operation and non-linear activation. Increasing the number of convolutional layers can enhance the feature extraction ability [35,36]. Multiple Res blocks are added after the convolutional layer while designing the network. Unlike the convolution unit, skip connection proposed by He et al. [37] is built into this block, which can effectively avoid gradient explosion or gradient disappearance, as well as increasing the training speed.
The size of the input and output of the convolutional layers keeps the same as the despeckling networks designed in [6,25,27], which increases the amount of computation to a certain extent. We reduce the amount of calculation by decreasing the dimension of the feature maps, i.e., adopting the pooling layer. We choose max pooling operation with the 2 × 2 pooling kernel in this layer. It should be noted that the pooling layer can also be replaced by strided convolutions [38].
The decoder consists of the convolutional layer and the sub-pixel units. The width and height of the input feature map to the decoder are only 1 / 4 of the original image after down-sampling twice through the pooling layer. Therefore, the up-sampling operation is required to make the output image of the network the same as the input size. However, an up-sampling operation such as transposed convolution used in [39,40] needs a high amount of computation and causes unwanted checkerboard artifacts [41,42]. A typical checkerboard pattern of artifacts is shown in Figure 4. To reduce the network runtime and avoid the checkerboard pattern of artifacts, the sub-pixel convolution described in Section 3.3 is used to implement the up-sampling operation.

3.3. Sub-Pixel Convolution

Sub-pixel convolution, also called as pixel shuffle, is an upscaling method first proposed in [43] for image super-resolution tasks. Different from the commonly used up-sampling methods in deep learning such as transposed convolution and fractionally strided convolution, sub-pixel convolution adopts channel to space method which achieves spatial scale-up amplification by rearranging pixels in multiple channels of the feature map, as illustrated in Figure 5.
For a sub-pixel unit with r times up-sampling, its output image is defined as I u p , and we have I u p R W × H × c , in which W, H, and c denote the width, height and channels of I u p . The sub-pixel convolution operation is defined as:
I u p ( x , y , c ) = F ( x / r , | y / r , c · r · mod ( y , r ) + c · mod ( x , r ) + c )
where I u p ( x , y , c ) is the value of the pixel at the position (x,y) for the c t h channel. F is the input of sub-pixel and F R W / r × H / r × c r 2 . · represents floor function that takes as input a real number and gives as output the greatest integer less than or equal to it [43]. After sub-pixel convolution operation, the elements of F are rearranged to the output I u p by increasing the horizontal and vertical count, and decreasing channel count. For example, when a 64 × 64 × 4 feature map is passed through the sub-pixel unit, an output with shape 128 × 128 × 1 will be obtained.

3.4. Proposed Evaluation Criterion

In this paper, the peak signal to noise ratio (PSNR) [44], structural similarity (SSIM) [45], equivalent number of looks (ENL) [46], and two new proposed evaluation criterions edge feature keep ratio (EFKR) and feature point keep ratio (FPKR) are used to evaluate the performance of despeckling methods.
PSNR is the ratio between the maximum possible power of a signal and the power of corrupting noise that affects the fidelity of its representation, which has been widely used in quality assessment of reconstructed images. SSIM is a metric of image similarity. ENL can describe the smoothness of regions, and no reference image is needed for its calculation, so it can be used to evaluate the performance of despeckling methods for real SAR images.

Edge Feature Keep Ratio and Feature Point Keep Ratio

PSNR and SSIM can effectively evaluate the overall performance of despeckling methods. Specifically, PSNR measures noise level or image distortion, SSIM measures the similarity between two images, and ENL measures the degree of region smoothing. They are not, however, capable of evaluating the edge and typical features retention ability in despeckling tasks directly. In this section, we propose two evaluation criteria that can compensate for the above deficiencies, i.e., edge feature keep ratio (EFKR) and feature point keep ratio (FPKR).
(a) EFKR: from the edge detection results shown in Figure 6, we have the following observations. (1) The edge outline of the speckled image is blurred, and there are discrete points in the image; (2) the edge outline is clear after despeckling and there is no discrete point, which is in agreement with the edge detection results of a clean image. Enlightened by this phenomenon, we design a quantitative evaluation criterion EFKR with the ability of edge retention based on counting the number of pixels of edges. The computation steps are as follows:
  • Edge detection processing for clean and test image using edge detection algorithms such as Sobel [47], Canny [48], Prewitt [49], and Roberts [50] methods. After this, two images only with only edge lines are obtained. The values of pixels in the edge position are set to 1, the values of the other position are set to 0.
  • Bit-wise and operation on two images from step 1, values at the position where the edges coincide are set to 1, values at other locations are set to 0.
  • Count the number of value 1 in the edge detected the result of a clean image and the number of 1 in step 2, and calculate the ratio of the two numbers.
The ratio of these two factors is the edge feature retention ratio, which is defined as:
E F K R = s u m ( edge ( X ) & edge ( Y ) ) s u m ( edge ( Y ) ) ,
where & and s u m denotes the bit-wise conjunction operation and sum operation, and edge represents edge detection.
(b) FPKR: for real SAR images, ENL is only able to evaluate the smooth level but not the retention ability of typical features such as edges, corners in the image. SIFT [51] can find feature points from different scales and obtain the ultimate descriptor of features. Also, the key points found by SIFT are usually corner points, edge points, bright spots in dark areas, and dark points in bright areas. These points are robust to light, affine transformations, and other transformation. The registration method based on SIFT first uses SIFT to obtain the feature points of the image to be registered, the reference image and their descriptor then matches the feature points according to descriptor and obtains one-to-one corresponding feature point pairs. Finally, the transformation parameters are calculated, and the image registration is carried out.
For SAR images, the registration of feature points and descriptors at the lights spots of speckles are redundant, which also reduce the efficiency and accuracy of subsequent searching of matching points. Based on this phenomenon, we design an evaluation criterion FPKR targeting at key feature points. We first execute an affine transformation to the evaluation image, then use the SIFT algorithm to find the feature points in the two images before and after the transformation, and finally match the feature points. The better the despeckling performs, the more typical features are preserved. The more prominent the feature descriptor obtained by SIFT, the greater the difference of descriptor between different features, so more effective feature point pairs can be searched efficiently. FPKR is defined as:
F P K R = N m a t c h ( X , X t ) min ( N ( X ) , N ( X t ) )
where N ( X ) , N ( X t ) are the number of key points before and after SIFT, and N m a t c h ( X , X t ) denotes the number of points for calculating transformation parameters.

4. Experiments and Results

4.1. Dataset

Because it is hard to collect real SAR images without speckle noise, we train networks by using synthetic noise/clean image pairs. Public dataset UC Merced Land Use Dataset (http://weegee.vision.ucmerced.edu/datasets/landuse.html) is chosen as the original clean image for training. The dataset contains 21 scene classes with 100 optical remote sensing images per class. Each image has a size of 256 × 256 pixels and the pixel resolution is 1 foot [52]. According to [27], we randomly select 400 images from the dataset as the training set and use the remaining images for testing. Some training samples are shown in Figure 7. Finally, after grayscale preprocessing, the speckled images are generated using Equation (1) same as the [25,53]. The noise levels (L = 2, 4, 8, 12) correspond to the number of looks in SAR, and the code of adding speckle noise is available on GitHub (https://github.com/rcouturier/ImageDenoisingwithDeepEncoderDecoder/tree/master/data{_}denoise).

4.2. Experimental Settings

All the networks are trained with stochastic gradient descent (SGD) with a mini-batch size of 32. All weights are initialized by a modified scheme of Xavier initialization [54] proposed by He et al. Meanwhile, we use the Adam optimizer [55] with tuned hyper-parameters to accelerate training. The hyper-parameters are kept the same across all layers and all networks. Experiments are implemented on TensorFlow platform with Intel i7-8700 CPU and an NVIDIA GTX-1080(8G) GPU.
The details of the model are specified here. The number of kernels in each unit is shown in Figure 3. The kernel sizes for the first and last convolutional layers are 5 × 5 , while all others are 3 × 3 . Rectified Linear Units (ReLU) are used as the activation function for all layers except for the last convolutional layer before sub-pixel unit. The L 1 loss is chosen to train the network, which is defined as:
L 1 ( Θ ) = 1 N y = 1 W x = 1 H | φ X ( x , y ) ; Θ C ( x , y ) | ,
where Θ is the filter parameters that need to be updated during training, C , X and φ · denote the objective image without noise, the input image with speckle noise, and the output after despeckling, respectively.

4.3. Experimental Results

The test results of our proposed network will be presented in this section. To verify the proposed method, we compare the performance of our MSR-net with other three despeckling methods, SAR-BM3D [22], ID-CNN [25], and Residual Encoder-Decoder network (RED-NET) [53]. The first one is a traditional nonlocal algorithm based on wavelet shrinkage, and the latter two methods are based on deep convolutional neural networks.

4.3.1. Results on Synthetic Images

Building, freeway, and airplane, three classes of synthetic images, are chosen as the test set to evaluate the noise reduction ability of each method. Part of the processing results of different algorithms under different levels of noise are shown in Figure 8 and Figure 9.
From the figures, we can observe that the CNN-based methods, including our MSR-net, can preserve more details like texture features in images than SAR-BM3D after despeckling. When the noise is strong, the SAR-BM3D algorithm will cause blurring at the edge of the objects.
ID-CNN has a good performance on image despeckling, however, after filtering by the network, pepper and salt noise appear in the image, which needs to be processed subsequently by using nonlinear filters such as median filtering and pseudo-median filtering. As the noise intensity increases, the salt and pepper noise increase gradually.
MSR-net has excellent retention performance of spatial geometry features like texture features, lines, and feature points. Compared with the other three algorithms, MSR-net has a higher smoothness of smooth areas as well as a smaller loss of sharpness of edges and details, especially for strong speckle noise. Also, more detail information in image will loose when the speckle noise is strong and more local detail can be preserved in output images when the speckle noise is weak.
When the level of noise added to the test set is small, all the CNN-based approaches can get state-of-art results. Therefore, it is difficult to judge the merits of these algorithms by using visual assessments. Experimental results of evaluation indexes such as PSNR and SSIM are necessary for these circumstances. The PSNR, SSIM, and EFKR evaluation indexes of the above methods are listed in Table 1, Table 2, Table 3 and Table 4, respectively. The bold number represents the optimal value in each row, while the underlined number denotes the suboptimal value. We also test MSR-net with only one scale and call it a single scale network (SS-net) during the experiment.
Consistent with the results shown in Figure 8 and Figure 9, our method has much better speckle-reduction ability than non-learned approach SAR-BM3D at different noise levels. In addition, the advantage of MSR-net will increase as the noise level increases. Taking airplane images for instance, the PSNR/SSIM/EFKR of our proposed MSR-net outperform SAR-BM3D by about 3.082 dB/0.047/0.2075, 2.785 dB/0.036/0.0895, 2.133 dB/0.022/0.0421, 1.944 dB/0.019/0.0425 for L = 2, 4, 8, 12.
Compared with CNN-based methods, MSR-net still has an advantage when the noise is strong. When L = 2, the PSNR/SSIM/EFKR of MSR-net outperform ID-CNN, RED-Net by about 1.129 dB/0.014/0.1434, 2.356 dB/0.144/0.1689, 0.427 dB/0.008/0.2152 and 0.447 dB/0.012/0.0191, 0.369 dB/0.018/-0.006, 0.643 dB/0.011/0.0194 for building, freeway, and airplane, respectively. When L = 4, the PSNR/SSIM/EFKR of MSR-net outperform ID-CNN, RED-Net by about 0.051 dB/0.134/0.0562, 0.885 dB/0.052/0.0589, 0.299 dB/0.016/0.0722 and 0.004 dB/-0.002/0.0375, 0.072 dB/0.004/0.0139, 0.197 dB/0.002/0.0143 for building, freeway and airplane, respectively.
In addition, we can see that our network does not always achieve the best test results. We consider this may have a certain relationship with the feature distribution of the images. Although MSR-net can only get sub-optimal test results for a certain class of images, the difference to the best result is small. For other classes of images, the advantages of MSR-net are more considerable. For example, we can observe from Table 1 that the EFKR of RED-Net only outperforms MSR-net by about 0.006 for freeway images, while MSR-net can outperform RED-Net by about 0.0191 and 0.0194 for building and airplane images. When the noise level L = 12, our network only gets four best test values, which suggests the advantage of the multi-scale network becomes smaller while the noise is weaker, as shown in Table 4.
MSR-net with a single scale (SS-net) also has very good speckle-reduction ability. When the noise intensity is weak, the performance is even better than multi-scale. For example, when L = 12, PSNR/SSIM/EFKR of the freeway are 28.787 dB/0.776/0.6247 and 28.893 dB/0.778/0.6197 for SS-net and MSR-net, respectively. Ultimately, we can find by comparison that the edge detection effect of the image has been significantly improved after despeckling.

4.3.2. Results on Real Sar Images

To further verify the speckle-reduction ability of our network for real SAR images, two SAR sceneries are selected, as shown in Figure 10a and Figure 11a, and these two images are imaging results of spaceborne SAR RADARSAT-2.
It can be seen by comparing the subgraphs in Figure 10 and Figure 11 that MSR-net generates the visually best output among all the results and retains the edge sharpness as well as the detail information about the structure in the image while removing the speckle noise. After filtering by SAR-BM3D, the loss of edge sharpness of the original SAR image is obvious and most of the lines and texture feature are blurred. ID-CNN and RED-Net can generate smooth results in homogeneous regions while maintaining textural features in the image. However, from the red boxes, we can observe that they are capable of retaining some texture features but not better than MSR-net. Although SS-net performs well in despeckling for real SAR images, it is still worse than multi-scale MSR-net.
The ENL results are shown in Table 5. We can observe from the table that MSR-net has an outstanding performance for real SAR image despeckling. For these four evaluation regions, the three highest scores and one second highest score of the ENL are obtained by our MSR-net.
To achieve FPKR results for real SAR data with different methods, we first apply the same affine transformation to each image, as shown in Figure 12. SIFT is then applied to search for the feature points and calculate their descriptors. Matching key points are ultimately conducted by minimizing the Euclidean distance between their SIFT descriptors. Generally, the ratio between distances is used [56] to obtain high matching accuracy. In the experiments, we select three ratios and the FPKR results are shown in Table 6 and Table 7. By comparing EFKR of each image, MSR-net performs better than SAR-BM3D. Also, MSR-net shows advantages over other neural network-based algorithms. Specifically, MSR-net achieves the best testing results in five out of six sets of experiments. It also indicates that pre-processing to SAR images by MSR-net can effectively enhance the usefulness of SIFT algorithm to SAR images and improve its performance and efficiency.

4.3.3. Runtime Comparisons

To evaluate the algorithm efficiency, we make statistics of the runtime of each algorithm in CPU implementation. The runtime of different methods on images with different sizes is listed in Table 8. We can see that the proposed denoiser is very competitive although its structure is relatively complex. Such a good compromise between speed and performance over MSR-net is properly attributed to the following two reasons. First, two pooling layers that can achieve spatial dimensionality reduction are embedded in the MSR-net. Each pooling layer with the 2 × 2 pooling kernel can reduce the amount of data that needs to be processed by the subsequent convolution operation to 25% of before. Second, in contrast to the transposed convolution which increases the resolution of feature maps by padding and complex convolution operation, sub-pixel unit, which up-samples feature maps by a periodic shuffling of pixel values, is adopted to build our network.

5. Discussion

5.1. Choice of Scales

To select a proper scale, building, freeway, and airplane, three classes of test images are used to analyze. Figure 13 shows the testing results of networks with different scales. In the single-scale network, recurrent modules ConvLSTM are replaced by a convolution layer to keep the same number of convolution layers.
We can observe that as the scale increases, the values of the three evaluation metrics are all improved. It suggests that exploiting the multi-scale information can help with improving network performance. However, we can meanwhile find that the improvement is small while the scale is greater than 3. We thus choose s=3 in our network to balance the network performance and complexity.

5.2. Loss Function

The influence of loss function on network performance is also discussed in this paper. Instead of L 1 loss, L 2 loss function is used to train our MSR-net. L 2 norm loss, also called Euclidean loss, is the most commonly used loss function in despeckling tasks. It is defined as:
L 2 ( Θ ) = 1 2 N y = 1 W x = 1 H φ X ( x , y ) ; Θ C ( x , y ) 2 2
where Θ is the filter parameter that needs to be updated during the training process, C is the ground truth image without noise, X is the input image with speckle noise, and φ · is the output after despeckling. The purpose of training network is to minimize the cost. Smaller loss value suggests a smaller error between the network output and its corresponding ground truth.
As shown in Figure 14, the network trained by L 2 loss function is more likely to obtain a higher PSNR only for building images and the network trained by L 1 loss function can obtain both slightly higher PSRN and SSIM with the other images. But for EFKR, the advantage of L 1 loss is significant compared to L 2 loss. Generally speaking, the L 1 loss is more suitable to SAR despeckling task.

6. Conclusions

In summary, different from the existing despeckling network, MSR-net proposed in this paper adopts the coarse-to-fine structure and the convolutional long short-term memory unit that can obtain high-quality despeckling SAR images. During research, we find that the weights sharing strategy of convolutional kernels can reduce network parameters and training complexity, and the sub-pixel unit used in this work can reduce up-sampling complexity, improve network efficiency, and shorten the runtime of the network with respect to the transposed convolutional layer. Meanwhile, new design evaluation metrics EFKR and FPKR are introduced herein to evaluate the compatibility of the despeckling algorithms to the optical image processing algorithms. Experimental results show that our MSR-net has excellent despeckling ability and achieves the state-of-the-art results both for simulated and real SAR images with low computational costs, especially in low signal noise ratio cases. The adaptability of optical image processing algorithms to SAR images can be enhanced after despeckling in our network.

Author Contributions

All of the authors made significant contributions to the work. Y.Z. and J.S. designed the research and analyzed the results. Y.Z. performed the experiments. Y.Z. and X.Y. wrote the paper. C.W., D.K., S.W., and X.Z. provided suggestions for the preparation and revision of the paper.

Acknowledgments

This work was supported in part by the Natural Science Fund of China under Grant 61671113 and the National Key R&D Program of China under Grant 2017YFB0502700, and in part by the Natural Science Fund of China under Grants 61501098, and 61571099.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ranjani, J.J.; Thiruvengadam, S. Dual-tree complex wavelet transform based SAR despeckling using interscale dependence. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2723–2731. [Google Scholar] [CrossRef]
  2. Ahishali, M.; Kiranyaz, S.; Ince, T.; Gabbouj, M. Dual and Single Polarized SAR Image Classification Using Compact Convolutional Neural Networks. Remote Sens. 2019, 11, 1340. [Google Scholar] [CrossRef]
  3. Jun, S.; Long, M.; Xiaoling, Z. Streaming BP for non-linear motion compensation SAR imaging based on GPU. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 6, 2035–2050. [Google Scholar] [CrossRef]
  4. Tang, X.; Zhang, X.; Shi, J.; Wei, S.; Tian, B. Ground Moving Target 2-D Velocity Estimation and Refocusing for Multichannel Maneuvering SAR with Fixed Acceleration. Sensors 2019, 19, 3695. [Google Scholar] [CrossRef]
  5. Lee, J.S.; Wen, J.H.; Ainsworth, T.L.; Chen, K.S.; Chen, A.J. Improved sigma filter for speckle filtering of SAR imagery. IEEE Trans. Geosci. Remote Sens. 2008, 47, 202–213. [Google Scholar]
  6. Yue, D.X.; Xu, F.; Jin, Y.Q. SAR despeckling neural network with logarithmic convolutional product model. Int. J. Remote Sens. 2018, 39, 7483–7505. [Google Scholar] [CrossRef]
  7. Xie, H.; Pierce, L.E.; Ulaby, F.T. Statistical properties of logarithmically transformed speckle. IEEE Trans. Geosci. Remote Sens. 2002, 40, 721–727. [Google Scholar] [CrossRef]
  8. Lee, J.S.; Hoppel, K.W.; Mango, S.A.; Miller, A.R. Intensity and phase statistics of multilook polarimetric and interferometric SAR imagery. IEEE Trans. Geosci. Remote Sens. 1994, 32, 1017–1028. [Google Scholar]
  9. Lattari, F.; Gonzalez Leon, B.; Asaro, F.; Rucci, A.; Prati, C.; Matteucci, M. Deep Learning for SAR Image Despeckling. Remote Sens. 2019, 11, 1532. [Google Scholar] [CrossRef]
  10. Lopes, A.; Touzi, R.; Nezry, E. Adaptive speckle filters and scene heterogeneity. IEEE Trans. Geosci. Remote Sens. 1990, 28, 992–1000. [Google Scholar] [CrossRef]
  11. Frost, V.S.; Stiles, J.A.; Shanmugan, K.S.; Holtzman, J.C. A model for radar images and its application to adaptive digital filtering of multiplicative noise. IEEE Trans. Pattern Anal. Mach. Intell. 1982, PAMI-4l, 157–166. [Google Scholar] [CrossRef]
  12. Kuan, D.T.; Sawchuk, A.A.; Strand, T.C.; Chavel, P. Adaptive noise smoothing filter for images with signal-dependent noise. IEEE Trans. Pattern Anal. Mach. Intell. 1985, PAMI-7, 165–177. [Google Scholar] [CrossRef]
  13. Lopes, A.; Nezry, E.; Touzi, R.; Laur, H. Structure detection and statistical adaptive speckle filtering in SAR images. Int. J. Remote Sens. 1993, 14, 1735–1758. [Google Scholar] [CrossRef]
  14. Kaur, L.; Gupta, S.; Chauhan, R. Image Denoising Using Wavelet Thresholding. ICVGIP Indian Conf. Comput. Vision Graph. Image Process. 2002, 2, 16–18. [Google Scholar]
  15. Cohen, A.; Daubechies, I.; Feauveau, J.C. Biorthogonal bases of compactly supported wavelets. Commun. Pure Appl. Math. 1992, 45, 485–560. [Google Scholar] [CrossRef]
  16. Donovan, G.C.; Geronimo, J.S.; Hardin, D.P.; Massopust, P.R. Construction of orthogonal wavelets using fractal interpolation functions. Siam J. Math. Anal. 1996, 27, 1158–1192. [Google Scholar] [CrossRef]
  17. Xie, H.; Pierce, L.E.; Ulaby, F.T. SAR speckle reduction using wavelet denoising and Markov random field modeling. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2196–2212. [Google Scholar] [CrossRef] [Green Version]
  18. Candes, E.J.; Donoho, D.L. Curvelets: A Surprisingly Effective Nonadaptive Representation for Objects with Edges; Technical Report; Stanford University, Department of Statistics: Stanford, CA, USA, 2000. [Google Scholar]
  19. Meyer, F.G.; Coifman, R.R. Brushlets: A tool for directional image analysis and image compression. Appl. Comput. Harmon. Anal. 1997, 4, 147–187. [Google Scholar] [CrossRef]
  20. Do, M.N.; Vetterli, M. The contourlet transform: an efficient directional multiresolution image representation. IEEE Trans. Image Process. 2005, 14, 2091–2106. [Google Scholar] [CrossRef] [Green Version]
  21. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  22. Parrilli, S.; Poderico, M.; Angelino, C.V.; Verdoliva, L. A nonlocal SAR image denoising algorithm based on LLMMSE wavelet shrinkage. IEEE Trans. Geosci. Remote Sens. 2011, 50, 606–616. [Google Scholar] [CrossRef]
  23. Deledalle, C.A.; Denis, L.; Tupin, F.; Reigber, A.; Jäger, M. NL-SAR: A unified nonlocal framework for resolution-preserving (Pol)(In) SAR denoising. IEEE Trans. Geosci. Remote Sens. 2014, 53, 2021–2038. [Google Scholar] [CrossRef]
  24. Kervrann, C.; Boulanger, J.; Coupé, P. Bayesian non-local means filter, image redundancy and adaptive dictionaries for noise removal. In Proceedings of the International Conference on Scale Space and Variational Methods in Computer Vision, Ischia, Italy, 30 May–2 June 2007; pp. 520–532. [Google Scholar]
  25. Wang, P.; Zhang, H.; Patel, V.M. SAR image despeckling using a convolutional neural network. IEEE Signal Process. Lett. 2017, 24, 1763–1767. [Google Scholar] [CrossRef]
  26. Bai, Y.C.; Zhang, S.; Chen, M.; Pu, Y.F.; Zhou, J.L. A Fractional Total Variational CNN Approach for SAR Image Despeckling. In Proceedings of the International Conference on Intelligent Computing, Bengaluru, India, 6 July 2018; pp. 431–442. [Google Scholar]
  27. Zhang, Q.; Yuan, Q.; Li, J.; Yang, Z.; Ma, X. Learning a dilated residual network for SAR image despeckling. Remote Sens. 2018, 10, 196. [Google Scholar] [CrossRef]
  28. Kaplan, L.M. Analysis of multiplicative speckle models for template-based SAR ATR. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 1424–1432. [Google Scholar] [CrossRef]
  29. Oliver, C.; Quegan, S. Understanding Synthetic Aperture Radar Images; SciTech Publishing: Noida, India, 2004. [Google Scholar]
  30. Wang, C.; Shi, J.; Yang, X.; Zhou, Y.; Wei, S.; Li, L.; Zhang, X. Geospatial Object Detection via Deconvolutional Region Proposal Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3014–3027. [Google Scholar] [CrossRef]
  31. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef]
  32. Xingjian, S.; Chen, Z.; Wang, H.; Yeung, D.Y.; Wong, W.K.; Woo, W.C. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2015; pp. 802–810. [Google Scholar]
  33. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  34. Chen, Q.; Koltun, V. Photographic image synthesis with cascaded refinement networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1511–1520. [Google Scholar]
  35. Srivastava, R.K.; Greff, K.; Schmidhuber, J. Highway networks. arXiv 2015, arXiv:1505.00387. [Google Scholar]
  36. He, K.; Sun, J. Convolutional neural networks at constrained time cost. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Santiago, Chile, 7–13 December 2015; pp. 5353–5360. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  38. Dumoulin, V.; Visin, F. A guide to convolution arithmetic for deep learning. arXiv 2016, arXiv:1603.07285. [Google Scholar]
  39. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Santiago, Chile, 7–13 December 2015; pp. 3431–3440. [Google Scholar]
  40. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  41. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar]
  42. Odena, A.; Dumoulin, V.; Olah, C. Deconvolution and checkerboard artifacts. Distill 2016, 1, e3. [Google Scholar] [CrossRef]
  43. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  44. Goodman, J.W. Statistical properties of laser speckle patterns. In Laser Speckle and Related Phenomena; Springer: Berlin/Heidelberg, Germany, 1975; pp. 9–75. [Google Scholar]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  46. Anfinsen, S.N.; Doulgeris, A.P.; Eltoft, T. Estimation of the equivalent number of looks in polarimetric synthetic aperture radar imagery. IEEE Trans. Geosci. Remote Sens. 2009, 47, 3795–3809. [Google Scholar] [CrossRef]
  47. Sobel, I. History and definition of the sobell operator. Retrieved World Wide Web. 2014. Available online: Available online: https://www.researchgate.net/publication/239398674 (accessed on 23 October 2019).
  48. Bao, P.; Zhang, L.; Wu, X. Canny edge detection enhancement by scale multiplication. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1485–1490. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  49. Dim, J.R.; Takamura, T. Alternative approach for satellite cloud classification: edge gradient application. Adv. Meteorol. 2013, 2013, 584816. [Google Scholar] [CrossRef]
  50. Davis, L.S. A survey of edge detection techniques. Comput. Graph. Image Process. 1975, 4, 248–270. [Google Scholar] [CrossRef]
  51. Tareen, S.A.K.; Saleem, Z. A comparative analysis of sift, surf, kaze, akaze, orb, and brisk. In Proceedings of the IEEE 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–10. [Google Scholar]
  52. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  53. Couturier, R.; Perrot, G.; Salomon, M. Image Denoising Using a Deep Encoder-Decoder Network with Skip Connections. In Proceedings of the International Conference on Neural Information Processing, Siem Reap, Cambodia, 13–16 December 2018; pp. 554–565. [Google Scholar]
  54. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  55. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  56. Kaplan, A.; Avraham, T.; Lindenbaum, M. Interpreting the ratio criterion for matching SIFT descriptors. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 697–712. [Google Scholar]
Figure 1. Inner structure of convolutional long short-term memory (ConvLSTM) [32].
Figure 1. Inner structure of convolutional long short-term memory (ConvLSTM) [32].
Remotesensing 11 02462 g001
Figure 2. The architecture of Multi-scale Recurrent. The yellow arrow denotes up-sampling by one time.
Figure 2. The architecture of Multi-scale Recurrent. The yellow arrow denotes up-sampling by one time.
Remotesensing 11 02462 g002
Figure 3. Architecture of Single Scale Modle
Figure 3. Architecture of Single Scale Modle
Remotesensing 11 02462 g003
Figure 4. Diagram of checkerboard artifacts. (a) original image without noise; (b) despeckling image with checkerboard artifacts caused by transposed convolution.
Figure 4. Diagram of checkerboard artifacts. (a) original image without noise; (b) despeckling image with checkerboard artifacts caused by transposed convolution.
Remotesensing 11 02462 g004
Figure 5. The sub-pixel convolutional operation on the input feature maps with an upscaling factor of r = 2, channel c = 1.
Figure 5. The sub-pixel convolutional operation on the input feature maps with an upscaling factor of r = 2, channel c = 1.
Remotesensing 11 02462 g005
Figure 6. Edge detection results. Images from left to right correspond to a clean image, an image with speckle noise, and image after despeckling.
Figure 6. Edge detection results. Images from left to right correspond to a clean image, an image with speckle noise, and image after despeckling.
Remotesensing 11 02462 g006
Figure 7. Part of the sample images used to train the network.
Figure 7. Part of the sample images used to train the network.
Remotesensing 11 02462 g007
Figure 8. Test results on images of airplanes with four levels of noise. The noise level from left to right is L = 2, L = 4, L = 8, and L = 12. (a) Speckled image (b) SAR-BM3D (Block-Matching and 3D filtering) (c) image despeckling convolutional neural network (ID-CNN) (d) RED-Net (e) multi-scale recurrent network (MSR-net).
Figure 8. Test results on images of airplanes with four levels of noise. The noise level from left to right is L = 2, L = 4, L = 8, and L = 12. (a) Speckled image (b) SAR-BM3D (Block-Matching and 3D filtering) (c) image despeckling convolutional neural network (ID-CNN) (d) RED-Net (e) multi-scale recurrent network (MSR-net).
Remotesensing 11 02462 g008
Figure 9. Test results on images of buildings with four levels of noise. The noise level from left to right is L = 2, L = 4, L = 8, and L = 12. (a) Speckled image (b) SAR-BM3D (c) ID-CNN (d) RED-Net (e) MSR-net.
Figure 9. Test results on images of buildings with four levels of noise. The noise level from left to right is L = 2, L = 4, L = 8, and L = 12. (a) Speckled image (b) SAR-BM3D (c) ID-CNN (d) RED-Net (e) MSR-net.
Remotesensing 11 02462 g009
Figure 10. Test results of Real SAR images (RADARSAT-2). (a) Original, (b) SAR-BM3D, (c) ID-CNN, (d) RED-Net, (e) single scale network (SS-net), (f) MSR-net.
Figure 10. Test results of Real SAR images (RADARSAT-2). (a) Original, (b) SAR-BM3D, (c) ID-CNN, (d) RED-Net, (e) single scale network (SS-net), (f) MSR-net.
Remotesensing 11 02462 g010
Figure 11. Test results of Real SAR images (RADARSAT-2). (a) Original, (b) SAR-BM3D, (c) ID-CNN, (d) RED-Net, (e) SS-net, (f) MSR-net.
Figure 11. Test results of Real SAR images (RADARSAT-2). (a) Original, (b) SAR-BM3D, (c) ID-CNN, (d) RED-Net, (e) SS-net, (f) MSR-net.
Remotesensing 11 02462 g011
Figure 12. Search results of feature point pairs for synthetic aperture radar (SAR) image before and after despeckling.
Figure 12. Search results of feature point pairs for synthetic aperture radar (SAR) image before and after despeckling.
Remotesensing 11 02462 g012
Figure 13. Test results of proposed network with scale 1, 2, 3, 4 when nosie level L = 2.
Figure 13. Test results of proposed network with scale 1, 2, 3, 4 when nosie level L = 2.
Remotesensing 11 02462 g013
Figure 14. Evaluation index (PSNR, SSIM, and EFKR) values of MSR-net with L1 and L2 Loss.
Figure 14. Evaluation index (PSNR, SSIM, and EFKR) values of MSR-net with L1 and L2 Loss.
Remotesensing 11 02462 g014
Table 1. The peak signal to noise ratio (PSNR), structural similarity (SSIM), and edge feature keep ratio (EFKR) of test set with noise level L = 2.
Table 1. The peak signal to noise ratio (PSNR), structural similarity (SSIM), and edge feature keep ratio (EFKR) of test set with noise level L = 2.
ClassesIndexSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
PSNR20.343 ± 0.04321.639 ± 0.07322.321 ± 0.06522.467 ± 0.03922.768 ± 0.052
building SSIM0.579± 0.0050.599 ± 0.0020.672 ± 0.0050.663 ± 0.0020.684 ± 0.004
EFKR0.3872± 0.0610.3888 ± 0.0580.5131 ± 0.0530.4950± 0.0400.5322 ± 0.049
PSNR22.478 ± 0.08123.17± 0.04625.157± 0.03325.255 ± 0.05725.526± 0.064
freeway SSIM0.581 ± 0.0020.536 ± 0.0030.662 ± 0.0030.663 ± 0.0480.680 ± 0.002
EFKR0.3046± 0.0520.3376 ± 0.0610.5071 ± 0.0610.4929 ± 0.0640.5065 ± 0.057
PSNR21.351± 0.04524.006 ± 0.05823.97± 0.06224.166 ± 0.04624.433± 0.049
airplane SSIM0.602 ± 0.0040.641 ± 0.0030.638 ± 0.0010.636 ± 0.0020.649 ± 0.005
EFKR0.3048± 0.0690.2965 ± 0.0730.4929 ± 0.0630.4800 ± 0.0580.5123 ± 0.065
The bold number represents the optimal value in each row, while the underlined number denotes the suboptimal value in each row.
Table 2. The PSNR, SSIM, and EFKR of test set with noise level L = 4.
Table 2. The PSNR, SSIM, and EFKR of test set with noise level L = 4.
ClassesIndexSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
PSNR21.789 ± 0.04624.319 ± 0.10124.366 ± 0.05924.279 ± 0.03224.370± 0.038
building SSIM0.694 ± 0.0030.721 ± 0.0050.737 ± 0.0060.722 ± 0.0010.735 ± 0.005
EFKR0.5029± 0.0590.5459± 0.0570.5646± 0.0500.5742± 0.04230.6021± 0.051
PSNR24.351 ± 0.05926.008 ± 0.05126.821 ± 0.06326.734 ± 0.03026.893 ± 0.049
freeway SSIM0.645 ± 0.0030.670 ± 0.0020.718 ± 0.0050.711 ± 0.0030.722 ± 0.003
EFKR0.4897± 0.0610.5019± 0.0600.5469± 0.0570.5511± 0.0650.5608± 0.051
PSNR23.261 ± 0.04125.747 ± 0.06725.849 ± 0.05025.905 ± 0.04626.046 ± 0.064
airplane SSIM0.657 ± 0.0030.677 ± 0.0050.691 ± 0.0040.686 ± 0.0020.693 ± 0.005
EFKR0.4976± 0.0610.5149 ± 0.0580.5484 ± 0.0560.5728 ± 0.0390.5871 ± 0.055
The bold number represents the optimal value in each row, while the underlined number denotes the suboptimal value in each row.
Table 3. The PSNR, SSIM, and EFKR of test set with the noise level L = 8.
Table 3. The PSNR, SSIM, and EFKR of test set with the noise level L = 8.
ClassesIndexSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
PSNR23.851 ± 0.07126.004 ± 0.05525.685 ± 0.04325.989 ± 0.06426.026 ± 0.048
building SSIM0.752 ± 0.0030.780 ± 0.0020.769 ± 0.050.774 ± 0.0040.782 ± 0.004
EFKR0.5992± 0.0650.6191± 0.0550.6305± 0.0500.6261± 0.05250.6476± 0.060
PSNR26.374 ± 0.08127.738 ± 0.04927.785 ± 0.06828.027 ± 0.05328.111 ± 0.057
freeway SSIM0.714 ± 0.0030.740 ± 0.0060.751 ± 0.0060.753 ± 0.0040.757 ± 0.003
EFKR0.5432± 0.0410.5809± 0.0530.5990± 0.0520.5890 ± 0.0570.5954± 0.048
PSNR25.31 ± 0.06427.438 ± 0.05127.161 ± 0.04427.404 ± 0.06127.443 ± 0.042
airplane SSIM0.71 ± 0.0050.731 ± 0.0040.720 ± 0.0040.728 ± 0.0050.732 ± 0.002
EFKR0.5872± 0.0570.6039 ± 0.0470.6113 ± 0.0530.6187 ± 0.0380.6293 ± 0.051
The bold number represents the optimal value in each row, while the underlined number denotes the suboptimal value in each row.
Table 4. The PSNR, SSIM, and EFKR of test set with the noise level L = 12.
Table 4. The PSNR, SSIM, and EFKR of test set with the noise level L = 12.
ClassesIndexSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
PSNR24.979 ± 0.06026.891 ± 0.04326.879 ± 0.05626.929 ± 0.07126.942 ± 0.052
building SSIM0.781 ± 0.0040.805 ± 0.00020.805 ± 0.0040.802 ± 0.0050.806 ± 0.003
EFKR0.6184± 0.0510.6441± 0.0490.6528± 0.0470.6596± 0.0540.6664± 0.058
PSNR27.444 ± 0.04228.643 ± 0.05928.905 ± 0.04728.787 ± 0.06628.893 ± 0.050
freeway SSIM0.747 ± 0.0030.768 ± 0.0060.782 ± 0.0070.776 ± 0.0040.778 ± 0.005
EFKR0.5975± 0.0560.6130± 0.0530.6218± 0.0560.6247± 0.0670.6197± 0.051
PSNR26.463 ± 0.05328.362 ± 0.6428.352 ± 0.07828.349 ± 0.05928.407 ± 0.065
airplane SSIM0.737 ± 0.0050.756 ± 0.0030.758 ± 0.0030.754 ± 0.0050.756 ± 0.006
EFKR0.6008± 0.0590.6275 ± 0.0570.6432 ± 0.0520.6465 ± 0.0520.6433 ± 0.053
The bold number represents the optimal value in each row, while the underlined number denotes the suboptimal value in each row.
Table 5. The equivalent number of looks (ENL) of real SAR regions
Table 5. The equivalent number of looks (ENL) of real SAR regions
DataOriginalSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
region1351104317394540
region2511951734132906644
region3380101333339622
region449995235369560
The bold number represents the optimal value in each row.
Table 6. Feature point keep ratio (FPKR) results before and after despeckling for real SAR image.
Table 6. Feature point keep ratio (FPKR) results before and after despeckling for real SAR image.
fp1fp2FPKR (r = 0.25)FPKR (r = 0.30)FPKR (r = 0.35)
original563745310.0004 ( 2 )0.0077 ( 35 )0.0362 (164)
SAR-BM3D455459470.0046 (21)0.0212 (110)0.0894 (407)
ID-CNN323339810.0049 (16)0.0294 ( 95 )0.1148 (371)
RED-Net363646440.0055 (20)0.0292 (106)0.0976 (355)
SS-Net342943570.0052 (18)0.0286 ( 98 )0.1155 (396)
MSR-net308042160.0071 (22)0.0425 (106)0.0981 (302)
The value inside brackets is the number of feature points. The bold number represents the optimal value in each column.
Table 7. FPKR results before and after despeckling for real SAR image.
Table 7. FPKR results before and after despeckling for real SAR image.
fp1fp2FPKR (r = 0.25)FPKR (r = 0.30)FPKR (r = 0.35)
original451134760.0012 ( 4 )0.0118 ( 41 )0.0437 (152)
SAR-BM3D387852890.0044 (17)0.0273 (106)0.0952 (369)
ID-CNN202125330.0124 (25)0.0574 (116)0.1366 (267)
RED-Net213727970.0140 (30)0.0543 (116)0.1526 (326)
MSR-net202125330.0166 (33)0.0674(134)0.1579 (314)
The value inside brackets is the number of feature points. The bold number represents the optimal value in each column.
Table 8. Runtime (s) of different methods of images with sizes of 256 × 256 and 512 × 512 .
Table 8. Runtime (s) of different methods of images with sizes of 256 × 256 and 512 × 512 .
SizeSAR-BM3DID-CNNRED-NetSS-NetMSR-Net
256 × 256 12.790.4482.2410.3390.586
512 × 512 52.8131.5628.9431.3422.375

Share and Cite

MDPI and ACS Style

Zhou, Y.; Shi, J.; Yang, X.; Wang, C.; Kumar, D.; Wei, S.; Zhang, X. Deep Multi-Scale Recurrent Network for Synthetic Aperture Radar Images Despeckling. Remote Sens. 2019, 11, 2462. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11212462

AMA Style

Zhou Y, Shi J, Yang X, Wang C, Kumar D, Wei S, Zhang X. Deep Multi-Scale Recurrent Network for Synthetic Aperture Radar Images Despeckling. Remote Sensing. 2019; 11(21):2462. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11212462

Chicago/Turabian Style

Zhou, Yuanyuan, Jun Shi, Xiaqing Yang, Chen Wang, Durga Kumar, Shunjun Wei, and Xiaoling Zhang. 2019. "Deep Multi-Scale Recurrent Network for Synthetic Aperture Radar Images Despeckling" Remote Sensing 11, no. 21: 2462. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11212462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop