Next Article in Journal
On-Orbit Autonomous Geometric Calibration of Directional Polarimetric Camera
Next Article in Special Issue
High-Performance Segmentation for Flood Mapping of HISEA-1 SAR Remote Sensing Images
Previous Article in Journal
Evaluation of Remote Sensing-Based Evapotranspiration Datasets for Improving Hydrological Model Simulation in Humid Region of East China
Previous Article in Special Issue
Fast Tree Detection and Counting on UAVs for Sequential Aerial Images with Generating Orthophoto Mosaicing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Semi-Supervised SAR ATR Framework with Transductive Auxiliary Segmentation

School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(18), 4547; https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184547
Submission received: 3 June 2022 / Revised: 8 September 2022 / Accepted: 8 September 2022 / Published: 12 September 2022
(This article belongs to the Special Issue Deep Learning in Remote Sensing Application)

Abstract

:
Convolutional neural networks (CNNs) have achieved high performance in synthetic aperture radar (SAR) automatic target recognition (ATR). However, the performance of CNNs depends heavily on a large amount of training data. The insufficiency of labeled training SAR images limits the recognition performance and even invalidates some ATR methods. Furthermore, under few labeled training data, many existing CNNs are even ineffective. To address these challenges, we propose a Semi-supervised SAR ATR Framework with transductive Auxiliary Segmentation (SFAS). The proposed framework focuses on exploiting the transductive generalization on available unlabeled samples with an auxiliary loss serving as a regularizer. Through auxiliary segmentation of unlabeled SAR samples and information residue loss (IRL) in training, the framework can employ the proposed training loop process and gradually exploit the information compilation of recognition and segmentation to construct a helpful inductive bias and achieve high performance. Experiments conducted on the MSTAR dataset have shown the effectiveness of our proposed SFAS for few-shot learning. The recognition performance of 94.18% can be achieved under 20 training samples in each class with simultaneous accurate segmentation results. Facing variances of EOCs, the recognition ratios are higher than 88.00% when 10 training samples each class.

1. Introduction

Synthetic aperture radar (SAR) is an important microwave remote sensing system in military and civil fields, and SAR automatic target recognition (ATR) is one of the most crucial and challenging issues in SAR applications. Recently, many outstanding scholars have proposed diverse deep learning-based methods and achieved remarkable results in SAR ATR applications [1,2,3,4,5].
However, many of these algorithms require that the training set consists of abundant labeled samples for each target type, which is often impossible to be satisfied in practical applications. Furthermore, under some conditions, e.g., earthquake rescue and sea rescue, the number of obtained SAR images can be only a few, and these existing SAR ATR methods probably lose their effectiveness. This has led to studies on few-shot learning (FSL) [6,7,8], which builds new classifiers from very few labeled SAR images, and these studies also continue to drive towards efficient and robust SAR ATR.
Previous approaches of FSL in SAR ATR mainly can be divided into three categories: data augmentation methods, metric-based methods, and model-based methods. The data augmentation methods improve the recognition performance by enriching the amount and enhancing the quality of training SAR images. For example, Zheng et al. proposed a semi-supervised SAR ATR method via a multi-discriminator generative adversarial network (GAN), which achieved an 85.23% recognition rate with 20 samples for each target [9]. Gao et al. proposed a semi-supervised GAN with multiple generators, which achieved more than 92% with around 40 samples in each target type [10].
The metric-based methods are based on learning representations for classes that can generalize to new classes. For example, Li et al. proposed a conv-biLSTM prototypical network, which has a classifier based on the Euclidean distance between training samples and the prototypical of each class [11]. Fu et al. designed a hard task mining method for effective meta-learning which yields better performance with the largest absolute improvements of 1.7% and 2.3% for 1-shot and 5-shot, respectively [12]. The model-based methods use prior knowledge to construct the embedding space and constrain the complexity. For example, Persello et al. proposed an active and semi-supervised learning network for SAR ATR, which is a semi-supervised support vector machine (SVM) [13]. Li et al. proposed a classification approach that adopts the canonical scattering models widely used in model-based decompositions to provide an improvement for the well-known H / α classification [14].
However, the challenge of FSL in SAR ATR is whether to have a helpful inductive bias [15,16,17], which can improve performance on novel classes but is hard to develop when the feature embedding has a large difference between a few samples and full samples [15,18]. In essence, these methods above just further mine the few labeled SAR images to obtain more usable recognition information. During the process, these methods cannot acquire new information from these few samples in terms of information theory, which hinders the improvement of recognition performance in SAR images.
In the light of deep segmentation methods for SAR images [19,20], the target segmentation has become more achievable and can reveal the unique morphological features that can be extended for use to SAR ATR under a few labeled training SAR samples. For example, Ref. [21] proposed a differentiable superpixel generation method and a superpixelwise statistical dissimilarity measure method with the encoder–decoder structure. Ref. [22] proposed one differentiable boundary-ware clustering method and the soft association map with an encoder–decoder structure. Meanwhile, the segmentation of SAR images is also beneficial for the SAR civil application [23,24]. For example, the segmentation of SAR images can help terrain mapping in debris flow rescue or urban construction design. Therefore, the segmentation of SAR images can be a basic and functional research field in SAR application.
To address the FSL challenge of SAR ATR, our key objective is to exploit the transductive generalization on unlabeled samples with an auxiliary loss serving as a regularizer. In doing so, not only the induction bias can be improved by narrowing the difference between a few samples and full samples, but also the transductive generalization can be exploited through an auxiliary loss. In this paper, we propose a semi-supervised SAR ATR framework with auxiliary segmentation (SFAS). This framework utilizes the core idea of transductive generalization to explore the relative features between unlabeled SAR samples and a few labeled SAR samples. Through auxiliary segmentation of unlabeled SAR samples and information residue loss (IRL) in training, the effectiveness of features extracted by the network is boosted, and the recognition of few labeled SAR samples is promoted greatly. In addition, accurate segmentation is also beneficial for SAR image interpretation, such as precision location or topographic mapping. The framework consists of three blocks: a shared feature extractor, a classifier to accomplish the recognition of labeled SAR target images, and a decoder to accomplish the segmentation of the unlabeled SAR target images. The contributions are as follows:
  • We focus on the transductive generalization on unlabeled samples and proposed a semi-supervised SAR ATR algorithm with an auxiliary segmentation loss as a regularizer, which improves the inductive bias of the model and provides a new source of abundant feature information for high recognition performance of SAR ATR FSL.
  • To explore the relative features between unlabeled SAR samples and few labeled SAR samples, we construct a training loop process (TCL) to alternately perform a semi-blind training optimization, and an information residue loss (IRL) to build a progressive optimization of recognition and segmentation. It can help the model optimize towards information compilation of segmentation and recognition.
  • The proposed framework achieves competitive performance to state-of-the-art methods on standard benchmarks and is easy to follow. The recognition rates of 40 training samples in each class are above 95.50%, and the rates of 20 training samples in each class are above 94.00%.
The remainder of this paper is organized as follows: the framework of the proposed SFAS and detailed information about some modules are presented in Section 2. Section 3 demonstrates the evaluation of performance based on the experimental results. Finally, Section 4 draws a brief conclusion.

2. Proposed Method

The motivation of the proposed algorithm is to utilize unlabeled samples which are relative to a few labeled SAR samples to explore the transductive generalization from unlabeled to few labeled. Limited by the practical circumstances, SAR ATR has to face the recognition of FSL, which leads to most of the existing methods failing to learn a helpful inductive bias on a few labeled samples. Nevertheless, in reality, there are often abundant unlabeled SAR samples which have a similar distribution with few labeled samples. By exploring the transductive generalization of unlabeled samples on an effective algorithm, the recognition of FSL in SAR ATR can be promoted and achieve state-of-the-art performance.
In this section, the proposed SFAS is introduced. First, we elucidate the training loop process and framework of the SFAS. Then, the information residue loss and specific baseline of the network are described in detail. To avoid confusion, it is specifically stated here that, in our paper, the labeled samples refer to the data with class labels but without segmentation labels, while unlabeled samples refer to the data with segmentation labels but without class labels.

2.1. Training Loop and Framework of Proposed SFAS

To explore the transductive generalization of unlabeled samples, we focus on extracting more generalized and effective features. Thus, we proposed a training loop process to make the network learn from multiple tasks and seek more generalized features which satisfy the optimization of multi-task. The description of the novel training loop is as follows.
As shown in Figure 1, during the training, there are mainly two steps: training data generating and training loop. In the training data generating phase, the random generator generates the unlabeled and labeled set required in the training process by sampling randomly from the training set. In addition, in one training loop, there is twice semi-blind training for recognition and segmentation.
We use D U to denote the unlabeled set, D L to denote the labeled set, M E , M D , M C to denote the extractor, decoder, and classifier, and P E , P D , and P C to denote the corresponding learnable parameters. In the first semi-blind training in one training loop, the unlabeled set D U goes through extractor M E and decoder M D to calculate a segmentation loss L S . The segmentation loss L S is only employed to update P D , and L S are combined with recognition loss L R of the last training loop based on IRL to update P E . Then, in the second semi-blind training, the few labeled set D L goes through extractor M E and decoder M C to calculate a segmentation loss L R . The loss L R is only employed to update P C , and L R are combined with L S based on IRL to update P E again.
By alternately performing semi-blind training for recognition and segmentation, this training loop can gradually exploit the transductive information of unlabeled samples to construct a helpful inductive bias on a few labeled samples.
Besides the shared feature extractor, a classifier and a decoder are shown in Figure 1, and the batch normalization and attention mechanism are employed in some level of multi-level feature extractors to ensure the network focus on the more effective features among the extracted features.

2.2. Information Residue Loss

The basic segmentation and recognition loss are set as the same form of cross-entropy cost. It should be noted that the design of the loss function for the classifier and decoder can promote each other and improve the accuracy simultaneously because the same form of loss can help construct the manifold structure and similar constraints. The details are as follows.
The recognition loss is set as the cross-entropy cost function, which is presented as
L r w , b = i = 1 C y i log p y i x ,
where p y i x is the probability vector of recognition result of the ith SAR chip, y i is the recognition labels, and C is the number of the recognition classes.
The function of the segmentation loss is defined as
L s w , b = 1 n i = 1 V s i log m s i x ,
where m s i x is the binary matrix of segmentation result of all pixel on the ith SAR chip, n is the number of pixels in a SAR chip, s i is the segmentation labels, and V is the number of the segmentation types.
Then, during the proposed training loop, the recognition loss is constructed based on the proposed information residue loss by combining the segmentation loss of the last training loop. The calculation is as follows.
For the recognition loss of IRL, it is defined as:
L r t = i = 1 C y i t log p t y i t x L , t α n i = 1 V s i t 1 log p t 1 s i t 1 x L , t 1 ,
where t is the training steps, and α denotes the parameter to adjust the residue ratio of segmentation information of the last loop.
The computation of the parameter α in IRL is
α = 1 t + 1 .
For the segmentation loss of IRL, the function is
L s t = 1 n i = 1 V s i t log p t s i t x L , t α i = 1 C y i t 1 log p t 1 y i t 1 x L , t 1
The proposed novel IRL utilizes the combination of the current recognition loss and information residue of previous segmentation loss to update the extractor or vice versa. This technology, at first, can lead to more effective optimization separately for recognition and segmentation. Then, when the training is going on, the integration of the losses of both recognition and segmentation can make the optimization of the extractor towards higher accuracy of the recognition and segmentation simultaneously, and further force the network gradually extract more generalized and effective features from unlabeled and labeled samples. It is helpful to exploit the transductive generalization from unlabeled samples to construct an effective inductive bias on a few labeled SAR samples.
Therefore, the proposed SFAS not only enables few-shot learning by employing unlabeled SAR target images but also provides an effective way to apply deep learning SAR ATR methods in practical applications by acquiring a small amount of labeled SAR images and various flexible SAR images as an auxiliary task.

2.3. Baseline Architecture

To exploit the transductive generalization on unlabeled samples with an auxiliary loss, each part of the architecture needs to be designed carefully. To extract more generalized and effective features for recognition and segmentation, the shared feature extractor needs to face the two challenges of recognition and segmentation in the training process. The classifier and decoder for the recognition and segmentation respectively need to employ the fine-tune of downstream subtasks.
Therefore, the baseline architecture is designed as shown in Figure 1. From the shared extractor in the blue box of Figure 1, the convolutional layers are linked by the residual way, and the attention mechanism block that contains spatial and channel attention modules [25] is employed to focus on the crucial local feature and compress the useless information in each layer. In addition, the leaky ReLu and batch normalization are employed to improve the generalization of the extractor.
The decoder shown in the green box of Figure 1 is mainly constructed by trans-convolutional blocks. These blocks consist of a trans-convolutional layer to reconstruct the structure of the SAR images, and a convolutional layer with attention and batch normalization to increase the generalization. Finally, a segmentation layer (Conv3-2) outputs the segmentation results. The classifier shown in the yellow box of Figure 1 is simply stacking the convolutional layer and max-pool layer to control the parameter of the classifier, which is an effective structure of classifier in SAR ATR.
Through the method described above, the recognition of SAR ATR under limited training data has the potential to achieve high performance. The next section shows the experiments and the results.

3. Experimental Results

In this section, the dataset for evaluation is presented. Then, we evaluated the performance of recognition and auxiliary segmentation under different sample numbers to validate the effectiveness and the transductive information of the proposed SFAS.

3.1. Dataset and Network Setup

The MSTAR dataset is a benchmark dataset for the SAR ATR performance assessment. The dataset contains a series of 0.3 m × 0.3 m SAR images of ten different classes of ground targets. The optical images and corresponding SAR images of ten classes of targets in the MSTAR dataset are shown in Figure 2. The training and testing data have the same ten target classes but different depression angles. The training data are captured under a depression angle of 17 , and the testing data are captured under that of 15 . It should be noted the ground truth of segmentation is roughly acquired by manual annotation using the tool named OpenLabeling, and there are also many available accurate automatic methods [19,20] to segment. The distribution of the training and testing images is listed in Table 1.
The hyperparameters in the proposed network are included in Table 2. The input SAR images have the size 80 × 80 . For each convolutional layer, the size of the stride is set as 1 × 1 . For each max-pooling layer, the size of the stride is set as 2 × 2 . The batch size is set as 64, and the learning rate is initialized as 0.0001. The labeled samples are based on the selected labeled training sample number in the experiments, and are randomly chosen from the whole dataset; then, the rest of the dataset is regarded as the unlabeled training samples. For example, if the number of selected labeled training samples is 60 for each class, it means that for each class, there are 60 randomly chosen labeled samples from the whole dataset, and the rest of the samples for each class are set as the unlabeled samples.
The proposed method is tested and evaluated on a computer with Inter Core I7-9700K at 3.6 GHz CPU, Gefore GTX 1080ti GPU with two 16 GB memories. The proposed method is implemented using the open-source PyTorch framework.

3.2. Recognition Performance under SOC

The recognition results of the proposed SFAS under standard operating condition (SOC) are shown in Table 3. From the table, facing 5-shot each target without pre-training, our method can achieve an 82.06% overall recognition ratio, and when the training number of each class in few labeled SAR samples is 20 and 10 respectively, the recognition rate just decreases from 94.18% to 89.28%. It indicates that the transductive generalization on unlabeled samples can improve the recognition performance when the labeled samples decrease. When the number of each class in few labeled SAR samples is 40, it gets above 95.5%, and eight classes of all targets are above 98.00%. It shows that the inductive bias constructed by the proposed SFAS is beneficial for the recognition of few labeled samples, though the two classes of the target are more sensitive to the decreased labeled samples. In addition, it is clear that, when the number of each class in few labeled SAR samples is larger than 80, the recognition rate can get above 98.5%. It demonstrates the effectiveness of the proposed SFAS in SAR ATR under few labeled SAR samples. When all the training samples are employed, the proposed method can achieve state-of-art performance in SAR ATR.

3.3. Recognition Performance under EOCs

In the practical application, there are more complex challenges in SAR ATR. For example, the training samples and testing samples are acquired under larger depression angle differences, which leads to obvious recognition failures. Therefore, in this subsection, the recognition performance under extended operating conditions (EOCs) such as the variances of the depression angle, target configuration, and version are evaluated for the robustness and effectiveness of the proposed SFAS.
Thanks to SAR images being sensitive to the depression angle, the experiments under a larger depression angle, EOC-D, are evaluated. The training and testing samples are listed in Table 4. The training samples are captured at 17 , and the testing samples are captured at 30 . The recognition performances under EOC-D are listed in Table 5. From Table 5, it is clear that our SFAS can achieve high performance under 100, 80 and 40 training samples for each class. Furthermore, when the training samples for each class are 20 and 10, respectively, the recognition performances of EOC-D are 94.61% and 90.18%. It has illustrated that our method can handle the few-shot effect under a large depression angle.
At the same time, the variances of target configuration and version, EOC-C and EOC-V, are evaluated for the effectiveness and practicability of our SFAS. The training and testing samples for EOC-C and EOC-V are listed in Table 4. The training samples are captured at 17 , and the testing samples for EOC-C and EOC-V are captured under 17 and 15 . The recognition performances for EOC-C and EOC-V are listed in Table 6 and Table 7.
For the recognition performance of EOC-C, it is obvious that the SFAS can achieve high performance, even facing few samples for each class. When the numbers of training samples are larger than 20 for each class, the overall recognition performances are higher than 94%. In addition, the recognition ratio of 20 in each class is 94.23% and the recognition ratio of 10 in each class is 88.99%. When the number of training samples is decreasing from 100 to 20, the recognition of EOC-C is still effective. It indicated that the SFAS faced the few-shot effect and showed certain robustness under EOC-C.
From the recognition results of EOC-V, the recognition ratios under a different number of training samples for each class are higher than 90.000%. When the number of training samples is decreasing from 100 to 10, the recognition ratio under EOC-V is decreasing slowly and smoothly, from 97.23% to 92.03%. It has illustrated that the proposed SFAS has shown its effectiveness and robustness under EOC-V.
From the recognition performance of EOCs, the effectiveness and practicability of the proposed SFAS have been evaluated. For further validation, the segmentation results are shown as follows.

3.4. Segmentation Performance

The overall accuracy of the segmentation can achieve 99.10% for the segmentation of target and background. The computation equation of segmentation accuracy is quoted from [26]. The partial qualitative segmentation result of the proposed SFAS is shown in Figure 3. From the three rows, it is clear that the segmentation results are accurate compared with the ground truth. A quantitative comparison regarding target accuracy, background accuracy, accuracy, and Intersection over Union (IoU) is listed in Table 8. In conclusion, the proposed SFAS can employ the transductive generalization from auxiliary segmentation to improve the recognition of few labeled samples, and utilize the recognition features to back feed the segmentation.

3.5. Comparison and Ablation

Regarding the different numbers of labeled training images, we compare our methods with other SAR ATR methods, including five traditional methods (PAC+SVM, ADaboost, LC-KSVD, DGM, and Gauss) and few-shot learning methods. Among these few-shot learning methods, there are three data augmentation methods (GAN-CNN, MGAN-CNN, and ARGN), one metric-based methods (TMDC-CNNs), and four model-based methods (DNN1, DNN2, CNN, and Semisupervised). When the number of labeled training images is 20, 40, 80, and all samples from each target, it can be seen that our proposed SFAS achieves a higher recognition rate than other methods in Table 9. Even if the number of each target in the training dataset is only 20 samples, the recognition rate of our method is above 94%, while other methods are mostly below 86%. Therefore, the results can not only validate the effectiveness of our proposed SFAS in the case of employing all training images in the dataset but also in the case of few-shot learning.
For further automation and practicality, we utilized visual salience and morphological methods to automatically segment the target from the background, called auto-seg labels, and we replace the manual segmentation labels with the auto-seg labels to validate the effectiveness of our method. When the labeled samples are 40 and 20, the recognition ratios are 95.15% and 93.77%. It is clear that our method still achieves high performance under few-shot learning with rough auto-seg labels, which illustrated the effectiveness and practicability of our method.
The ablation experiments are also conducted shown in Table 10. When the labeled samples are 100, 80, 40, and 20, respectively, the recognition ratios without auxiliary segmentation loss are 98.93%, 95.09%, 84.62%, and 71.54% as shown in Table 10. The comparison illustrates that the segmentation loss is more effective facing few training samples.

3.6. Recognition Performance with Automatic Segmentation

In practice, samples with recognition labels may be impossible to obtain, and the segmentation label can be used to assist in recognizing small samples without recognition labels. Therefore, segmentation labels are helpful for recognition. If they are marked manually, it is really laborious, but some methods can automatically segment SAR images, which can also achieve recognition. We also conducted experiments to verify the above conclusions. We utilized visual salience and morphological methods to automatically segment the target from the background, called auto-seg labels. Then, we replace the manual segmentation labels with the auto-seg labels to validate the effectiveness of our method.
The recognition result of the proposed SFAS is shown in Table 11. The recognition ratio under a different number of training samples for each class is decreasing from 98.93% to 84.70%. When facing 100, 80 and 40 in each class, the recognition ratio achieved 98.93%, 95.67%, and 95.38%, respectively. It is clear that, with rough automatic segmentation, our SFAS still has the effectiveness and robustness of facing the few-shot effect. When the training samples for each class are 20 and 10, the overall recognition ratios with auto-segmentation are 93.19% and 84.70%. At the same time, the recognition ratios with manual segmentation are 94.18% and 89.28%. It showed that our SFAS still achieved good recognition performance and the precise segmentation can benefit the recognition. For the limited training samples, like 5-shot or 10-shot, our SFAS can employ precise segmentation from more effective segmentation methods for higher recognition ratios.
From the experiments with automatic segmentation, the practicability and effectiveness of our SFAS have been demonstrated. Even facing rough automatic segmentation, our SFAS still achieved high recognition performance and showed the robustness of the few-shot effect.

4. Conclusions

When the distributions of the few samples and complete samples have a large difference, the FSL of SAR ATR has failed to construct a helpful inductive bias. To tackle this challenge, we proposed a semi-supervised SAR ATR algorithm by employing a multi-level feature framework with auxiliary segmentation. By employing the supervised information of a few labeled samples and the transductive manifold of the unlabeled samples, the proposed novel SFAS exploits the transductive generalization to achieve better recognition performance of FSL in SAR ATR. Experimental results on the MSTAR dataset have validated the effectiveness and robustness of the proposed SFAS in few-shot recognition in SAR. The proposed SFAS not only can achieve the state of the art in recognition but also achieve superior performance in auxiliary segmentation.

Author Contributions

Conceptualization, C.W. and X.L.; methodology, C.W.; software, C.W.; validation, C.W., X.L. and Y.H.; formal analysis, Y.H. and J.P.; investigation, S.L., J.Y. and D.M.; resources, J.Y.; data curation, C.W. and X.L.; writing—original draft preparation, C.W. and X.L.; writing—review and editing, Y.H., J.P. and S.L.; visualization, C.W. and X.L.; supervision, J.Y. and D.M.; project administration, Y.H., J.P., J.Y. and D.M.; funding acquisition, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grants 61901091 and 61901090.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Geng, Z.; Yan, H.; Zhang, J.; Zhu, D. Deep-Learning for Radar: A Survey. IEEE Access 2021, 9, 141800–141818. [Google Scholar] [CrossRef]
  2. Kechagias-Stamatis, O.; Aouf, N. Automatic Target Recognition on Synthetic Aperture Radar Imagery: A Survey. IEEE Aerosp. Electron. Syst. Mag. 2021, 36, 56–81. [Google Scholar] [CrossRef]
  3. Oveis, A.H.; Guisti, E.; Ghio, S.; Martorella, M. A Survey on the Applications of Convolutional Neural Networks for Synthetic Aperture Radar: Recent Advances. IEEE Aerosp. Electron. Syst. Mag. 2021, 37, 18–42. [Google Scholar] [CrossRef]
  4. Kechagias-Stamatis, O.; Aouf, N. Fusing Deep Learning and Sparse Coding for SAR ATR. IEEE Trans. Aerosp. Electron. Syst. 2019, 55, 785–797. [Google Scholar] [CrossRef]
  5. Ding, B.; Wen, G. Target Reconstruction Based on 3D Scattering Center Model for Robust SAR ATR. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3772–3785. [Google Scholar] [CrossRef]
  6. Sung, F.; Yang, Y.; Zhang, L.; Xiang, T.; Torr, P.H.; Hospedales, T.M. Learning to Compare: Relation Network for Few-Shot Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  7. Wang, Y.; Yao, Q.; Kwok, J.T.; Ni, L.M. Generalizing from a Few Examples: A Survey on Few-Shot Learning. ACM Comput. Surv. 2020, 53. [Google Scholar] [CrossRef]
  8. Wang, L.; Bai, X.; Xue, R.; Zhou, F. Few-shot SAR automatic target recognition based on Conv-BiLSTM prototypical network. Neurocomputing 2021, 443, 235–246. [Google Scholar] [CrossRef]
  9. Zheng, C.; Jiang, X.; Liu, X. Semi-supervised SAR ATR via multi-discriminator generative adversarial network. IEEE Sens. J. 2019, 19, 7525–7533. [Google Scholar] [CrossRef]
  10. Gao, F.; Yang, Y.; Wang, J.; Sun, J.; Yang, E.; Zhou, H. A deep convolutional generative adversarial networks (DCGANs)-based semi-supervised method for object recognition in synthetic aperture radar (SAR) images. Remote Sens. 2018, 10, 846. [Google Scholar] [CrossRef]
  11. Wang, L.; Bai, X.; Zhou, F. Few-Shot SAR ATR Based on Conv-BiLSTM Prototypical Networks. In Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 26–29 November 2019; pp. 1–5. [Google Scholar]
  12. Fu, K.; Zhang, T.; Zhang, Y.; Wang, Z.; Sun, X. Few-Shot SAR Target Classification via Metalearning. IEEE Trans. Geosci. Remote Sens. 2021, 60, 2000314. [Google Scholar] [CrossRef]
  13. Persello, C.; Bruzzone, L. Active and semisupervised learning for the classification of remote sensing images. IEEE Trans. Geosci. Remote Sens. 2014, 52, 6937–6956. [Google Scholar] [CrossRef]
  14. Li, D.; Zhang, Y. Adaptive Model-Based Classification of PolSAR Data. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6940–6955. [Google Scholar] [CrossRef]
  15. Chen, Z.; Maji, S.; Learned-Miller, E. Shot in the dark: Few-shot learning with no base-class labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2668–2677. [Google Scholar]
  16. Hou, Z.; Kung, S.Y. Few-Shot Learning Via Dependency Maximization and Instance Discriminant Analysis. In Proceedings of the 2021 IEEE 31st International Workshop on Machine Learning for Signal Processing (MLSP), Gold Coast, Australia, 25–28 October 2021; pp. 1–6. [Google Scholar] [CrossRef]
  17. Ye, H.J.; Sheng, X.R.; Zhan, D.C. Few-shot learning with adaptively initialized task optimizer: A practical meta-learning approach. Mach. Learn. 2020, 109, 643–664. [Google Scholar] [CrossRef]
  18. Chen, Y.; Liu, Z.; Xu, H.; Darrell, T.; Wang, X. Meta-Baseline: Exploring Simple Meta-Learning for Few-Shot Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 9062–9071. [Google Scholar]
  19. Zhang, P.; Li, M.; Wu, Y.; Li, H. Hierarchical conditional random fields model for semisupervised SAR image segmentation. IEEE Trans. Geosci. Remote Sens. 2015, 53, 4933–4951. [Google Scholar] [CrossRef]
  20. Jing, H.; Sun, X.; Wang, Z.; Chen, K.; Diao, W.; Fu, K. Fine building segmentation in high-resolution SAR images via selective pyramid dilated network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6608–6623. [Google Scholar] [CrossRef]
  21. Ma, F.; Zhang, F.; Xiang, D.; Yin, Q.; Zhou, Y. Fast Task-Specific Region Merging for SAR Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  22. Ma, F.; Zhang, F.; Yin, Q.; Xiang, D.; Zhou, Y. Fast SAR Image Segmentation With Deep Task-Specific Superpixel Sampling and Soft Graph Convolution. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–16. [Google Scholar] [CrossRef]
  23. Jing, W.; Jin, T.; Xiang, D. Fast Superpixel-Based Clustering Algorithm for SAR Image Segmentation. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  24. Wan, L.; Zhang, T.; Xiang, Y.; You, H. A Robust Fuzzy C-Means Algorithm Based on Bayesian Nonlocal Spatial Information for SAR Image Segmentation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 896–906. [Google Scholar] [CrossRef]
  25. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  26. Wang, C.; Pei, J.; Wang, Z.; Huang, Y.; Wu, J.; Yang, H.; Yang, J. When Deep Learning Meets Multi-Task Learning in SAR ATR: Simultaneous Target Recognition and Segmentation. Remote Sens. 2020, 12, 3863. [Google Scholar] [CrossRef]
  27. Srinivas, U.; Monga, V.; Raj, R.G. SAR automatic target recognition using discriminative graphical models. IEEE Trans. Aerosp. Electron. Syst. 2014, 50, 591–606. [Google Scholar] [CrossRef]
  28. O’Sullivan, J.A.; DeVore, M.D.; Kedia, V.; Miller, M.I. SAR ATR performance using a conditionally Gaussian model. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 91–108. [Google Scholar] [CrossRef]
  29. Morgan, D.A. Deep convolutional neural networks for ATR from SAR imagery. In Algorithms for Synthetic Aperture Radar Imagery XXII; International Society for Optics and Photonics: Bellingham, WA, USA, 2015; Volume 9475, p. 94750F. [Google Scholar]
  30. Ding, J.; Chen, B.; Liu, H.; Huang, M. Convolutional neural network with data augmentation for SAR target recognition. IEEE Geosci. Remote Sens. Lett. 2016, 13, 364–368. [Google Scholar] [CrossRef]
  31. Guan, J.; Liu, J.; Feng, P.; Wang, W. Multiscale Deep Neural Network with Two-Stage Loss for SAR Target Recognition With Small Training Set. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  32. Sun, Y.; Wang, Y.; Liu, H.; Wang, N.; Wang, J. SAR Target Recognition With Limited Training Data Based on Angular Rotation Generative Network. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1928–1932. [Google Scholar] [CrossRef]
  33. Wang, C.; Shi, J.; Zhou, Y.; Yang, X.; Zhou, Z.; Wei, S.; Zhang, X. Semisupervised Learning-Based SAR ATR via Self-Consistent Augmentation. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4862–4873. [Google Scholar] [CrossRef]
Figure 1. Baseline architecture.
Figure 1. Baseline architecture.
Remotesensing 14 04547 g001
Figure 2. Optical images and corresponding SAR images of ten classes of objects in the MSTAR database. (From left to right: BMP2, BTR70, T72, 2S1, BRDM2, ZSU234, BTR60, D7, T62, and ZIL131).
Figure 2. Optical images and corresponding SAR images of ten classes of objects in the MSTAR database. (From left to right: BMP2, BTR70, T72, 2S1, BRDM2, ZSU234, BTR60, D7, T62, and ZIL131).
Remotesensing 14 04547 g002
Figure 3. Three randomly selected samples to show the segmentation performance. Each row represents one sample including original image, ground truth, and segmentation result (from left to right).
Figure 3. Three randomly selected samples to show the segmentation performance. Each row represents one sample including original image, ground truth, and segmentation result (from left to right).
Remotesensing 14 04547 g003
Table 1. MSTAR dataset used in experiments.
Table 1. MSTAR dataset used in experiments.
Target TypeBMP2BRDM2BTR60BTR70D72S1T62T72ZIL131ZSU235
Training (17 )233298256233299299299232299299
Testing (15 )195274195196274274273196274274
Table 2. Hyper-parameters of the proposed framework.
Table 2. Hyper-parameters of the proposed framework.
ModuleLayersBatchNorm or Attention
ExtractorConv3×3@16BatchNorm
ReLu + MaxPoolNeither
Conv3×3@32BatchNorm
ReLU + MaxPoolNeither
Conv3×3@64Both
ReLU + MaxPoolNeither
DecoderTrans-Conv3×3@64BatchNorm
Conv3×3@32Both
Trans-Conv3×3@32BatchNorm
Conv3×3@16Both
Trans-Conv3×3@16BatchNorm
Conv3×3@8Neither
Conv3×3@2Neither
ClassifierConv4×4@128BatchNorm
ReLU + MaxPoolNeither
Conv4×4@256BatchNorm
ReLU + MaxPoolNeither
SoftMaxNeither
Table 3. Recognition accuracy (%) of SOC under different numbers of labeled images in each classes.
Table 3. Recognition accuracy (%) of SOC under different numbers of labeled images in each classes.
ClassLabeled Number in Each Class
510204080100All
BMP238.9792.3198.4698.9797.9599.49100.00
BTR7093.8098.1897.45100.0099.4998.47100.00
T7271.7973.8596.4199.4995.92100.0099.49
BTR60100.0098.4796.4194.3698.4698.9797.44
2S198.9186.5093.8081.3998.5498.9197.81
BRDM267.5293.07100.0099.64100.0098.91100.00
D784.6286.0899.6498.1899.64100.0099.27
T6293.3783.6780.2489.0198.5399.64100.00
ZIL13186.8693.4391.8498.5497.4597.81100.00
ZSU23477.7484.6790.1599.6498.54100.00100.00
Average82.0689.2894.1895.6298.5299.2299.42
Table 4. Training and testing dataset under EOCs.
Table 4. Training and testing dataset under EOCs.
TrainNumberTest (EOC-D)Number
2S12992S1(b01)288
BRDM2298BRDM2(E71)287
T72232T72(A64)288
ZSU234299ZSU234(d08)288
TrainNumberTest (EOC-C)Number
BMP2233T72(S7)419
BRDM2298T72(A32)572
BTR70233T72(A62)573
T72232T72(A63)573
T72(A64)573
TrainNumberTest (EOC-V)Number
BMP2233T72(SN812)426
T72(A04)573
BRDM2298T72(A05)573
T72(A07)573
BTR70233T72(A10)567
BMP2(9566)428
T72232BMP2(C21)429
Table 5. Recognition accuracy (%) of EOC-D under different numbers of labeled images in each class.
Table 5. Recognition accuracy (%) of EOC-D under different numbers of labeled images in each class.
EOC-D
ClassLabeled Number in Each Class
10080402010
BRDM299.6599.3099.6599.65100.00
2S198.9698.6198.6199.6599.31
T72-A6492.7192.3687.8583.3373.26
ZSU23498.6197.2298.6195.8388.19
Average97.4896.8796.1894.6190.18
Table 6. Recognition accuracy (%) of EOC-C under different numbers of labeled images in each class.
Table 6. Recognition accuracy (%) of EOC-C under different numbers of labeled images in each class.
EOC-C
ClassLabeled Number in Each Class
10080402010
T72-A3297.7397.7396.8596.1595.99
T72-A6298.9597.7397.0394.9492.32
T72-A6397.7395.8193.5493.5492.15
T72-A6493.1993.1992.5091.9786.56
T72sn-s799.0597.1496.9094.5193.32
Average97.2396.2795.2894.2192.03
Table 7. Recognition accuracy (%) of EOC-V under different numbers of labeled images in each class.
Table 7. Recognition accuracy (%) of EOC-V under different numbers of labeled images in each class.
EOC-V
ClassLabeled Number in Each Class
10080402010
BMP2sn-956690.6591.3689.4963.0858.18
BMP2sn-c2190.9190.6890.2184.3856.41
T72sn-81298.3698.8398.5999.7799.30
T72-A0498.0897.0398.0897.7399.30
T72-A0596.6895.9996.6895.6498.78
T72-A0797.5697.0397.0394.4298.95
T72-A1098.7798.5998.5998.2498.77
Average96.1695.8895.8594.2388.99
Table 8. Segmentation performance (%) of different methods.
Table 8. Segmentation performance (%) of different methods.
MethodsTarget AccuracyBackground AccuracyAccuracyIoU
Otsu58.17%88.35%73.26%52.10%
Canny79.12%90.13%84.62%72.01%
Ours99.24%98.97%99.10%98.23%
Table 9. Recognition accuracy (%) under different numbers of labeled images.
Table 9. Recognition accuracy (%) under different numbers of labeled images.
MethodLabeled Number
ALL10080403020
PCA+SVM [27]94.32-92.4887.95-76.43
ADaboost [9]93.51-91.4586.45-75.68
LC-KSVD [9]95.13-93.2387.39-78.83
DGM [9]96.07-92.8588.14-81.11
Gauss [28]97.16-94.1087.51-80.55
DNN1 [29]95.54-93.0486.98-77.86
DNN2 [30]96.50-93.7687.73-79.39
CNN [9]97.03-93.8888.35-81.80
GAN-CNN [9]97.53-94.9190.13-84.39
MGAN-CNN [9]97.81-94.9190.82-85.23
TMDC-CNNs [31]-99.09--86.93-
ARGN [32]98.00-95.98-87.20-
Semisupervised [33]--98.6597.11-92.62
The Proposed99.42-98.5295.62-94.18
Table 10. Ablation experiments of segmentation labels.
Table 10. Ablation experiments of segmentation labels.
Sample Number10204080100
With segmentation loss89.2894.1895.6298.5299.22
Without segmentation loss62.5771.5484.6295.0998.93
Table 11. Recognition accuracy (%) with automatic segmentation.
Table 11. Recognition accuracy (%) with automatic segmentation.
ClassLabeled Number in Each Class
10204080100
BMP261.5492.3182.5690.2697.95
BTR7066.0684.6795.6288.6998.91
T7275.3890.2696.4192.3196.92
BTR6071.4388.7889.2997.4599.49
2S195.6290.8894.1693.0799.27
BRDM294.5397.0892.7099.2798.91
D784.2598.53100.0097.0798.17
T6297.4597.9699.4999.49100.00
ZIL13193.4395.99100.0098.1899.27
ZSU23497.8194.5399.64100.00100.00
Average84.7093.1995.3895.6798.93
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, C.; Liu, X.; Huang, Y.; Luo, S.; Pei, J.; Yang, J.; Mao, D. Semi-Supervised SAR ATR Framework with Transductive Auxiliary Segmentation. Remote Sens. 2022, 14, 4547. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184547

AMA Style

Wang C, Liu X, Huang Y, Luo S, Pei J, Yang J, Mao D. Semi-Supervised SAR ATR Framework with Transductive Auxiliary Segmentation. Remote Sensing. 2022; 14(18):4547. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184547

Chicago/Turabian Style

Wang, Chenwei, Xiaoyu Liu, Yulin Huang, Siyi Luo, Jifang Pei, Jianyu Yang, and Deqing Mao. 2022. "Semi-Supervised SAR ATR Framework with Transductive Auxiliary Segmentation" Remote Sensing 14, no. 18: 4547. https://0-doi-org.brum.beds.ac.uk/10.3390/rs14184547

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop