Multi-Aspect SAR Target Recognition Based on Prototypical Network with a Small Number of Training Samples
<p>Comparison of the same target’s SAR images in different azimuth angles. Sub-figures (1)–(6) are SAR images of T72 tank in different azimuth angles.</p> "> Figure 2
<p>Schematic diagram of method training.</p> "> Figure 3
<p>Multi-task learning model.</p> "> Figure 4
<p>Schematic diagram of NLECA module.</p> "> Figure 5
<p>Conv1d and Non-Local Conv1d.</p> "> Figure 6
<p>Non-local one-dimensional convolution structure diagram.</p> "> Figure 7
<p>Multi-channel EfficientNet-B0 embedded with NLECA module.</p> "> Figure 8
<p>Schematic diagram of multi-level feature extraction and fusion.</p> "> Figure 9
<p>Schematic diagram of model training process.</p> "> Figure 10
<p>Schematic diagram of model test.</p> "> Figure 11
<p>Diagram of recognition accuracy of ablation experiments.</p> "> Figure 12
<p>t-SNE visualization of testing dataset output.</p> ">
Abstract
:1. Introduction
- SAR image sequence with multiple azimuth angles can make up for the lack of information in a single image, and further increase the information for target recognition.
- There is a certain correlation involved in multi-aspect SAR images, which provides another dimensional of information for target recognition.
- Experiments show that this method can significantly improve the recognition performance of the model with a small number of training samples.
- This method can be applied to different deep learning feature extraction models.
2. Methods
2.1. Prototypical Network
2.2. Multi-Task Learning
2.3. Feature Extraction Model
2.3.1. EfficientNet
2.3.2. NLECA Module
2.3.3. NLECA-EfficinentNet
2.4. Multi-Level Feature Fusion
2.5. Training and Testing Tricks
2.5.1. Image Preprocessing
2.5.2. Label Smoothing
2.5.3. Test Time Augmentation
2.6. Multi-Aspect SAR ATR Method with Small Number of Training Samples
- Randomly select samples from each of the N classes in the training set as the support set , and select samples as the query set , where is the multi-aspect SAR images, is the sample label, and there are a total of training samples participate in each training episode.
- Use the training samples in the support set to calculate the prototype of each class.That is, the support set samples of each class are used to obtain multi-level feature vectors through the multi-aspect SAR feature extraction model, and the multi-level feature vectors of each class are averaged to obtain the prototype.
- The samples of the query set are used to obtain the feature vector through the feature extraction model, which is used to calculate the Euclidean distance from each prototype. The loss is calculated using Formulas (2) and (4).
- Use the multi-level feature vectors of all samples in the support set and the query set to perform the classification task, and obtain the classification loss by Formula (8).
- Perform a weighted addition on the prototype loss and the classification loss to get the total loss, while backpropagates to adjust the model parameters.
- Use the trained feature extraction model and use the entire training set to calculate the prototype of each class.
- Rotate each image of the test sample by , and extend the test sample by three times, obtain three feature vectors by the trained feature extraction model, and average the three feature vectors to obtain a new feature vector.
- Use the feature vector in step 2 and the prototype in step 1 to calculate the Euclidean distance and obtain the distance vector, and then use the Softmax classifier to classify the distance vector.
3. Experiments and Results
3.1. Experimental Dataset
3.1.1. MSTAR Dataset
3.1.2. Small Number of Training Datasets
- SOC training set: This paper establishes two under-sampled SOC datasets, SOC-150 and SOC-250. SOC-150 has 10 classes with only 15 samples per class, a total of 150 training samples; SOC-250 has 10 classes with only 25 samples per class, a total of 250 training samples. The two datasets only account for 1.42% and 2.36% of the original training set size, respectively.
- EOC1 training set: After under-sampled from original EOC1 training set, two datasets, EOC1-100 and EOC1-200, are obtained. There are only 25 samples in each of the four classes of EOC1-100, a total of 100 training samples.There are only 50 samples in each of the 4 classes in EOC1-200, a total of 200 training samples. The two datasets account for 2.26% and 4.53% of the original training set size, respectively.
- EOC2 and EOC3 training sets: After under-sampled from original EOC2 training set (the same as original EOC3 training set), two datasets, EOC23-100 and EOC23-200 are obtained. The EOC23-100 data set has 4 classes with only 25 samples per class, a total of 100 training samples. There are 50 samples in each of the 4 classes in E0C23-200 dataset, a total of 200 training samples. The two datasets account for 2.23% and 4.47% of the original training set size, respectively.
3.2. Experimental Environment
3.3. Experimental Parameters
- Use the method proposed in this paper for training and testing. The Epochs for training are all set to 100, where each epoch trains 200 episodes, the Adam optimizer is used for learning, the learning rate is 1e-3, and the learning rate for each 20 epoch is reduced to 1/2. For SOC-150 training set, the number of support sets for each episode is 5, and the number of query sets is 10. For SOC-250 training set, the number of support sets is 10, and the number of query sets is 15. For the training sets of the EOC1 (EOC1-100, EOC1-200) and EOC23 (EOC23-100, EOC23-200), the number of support sets for each episode is 10, and the number of query sets is 10.
3.4. Results of NLECA-EfficientNet Experiments
3.5. Results of Ablation Experiments
4. Discussion
4.1. Advantages
4.2. Generality
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Bhanu, B. Automatic target recognition: State of the art survey. IEEE Trans. Aerosp. Electron. Syst. 1986, 4, 364–379. [Google Scholar] [CrossRef]
- Wang, P.; Liu, W.; Chen, J.; Niu, M.; Yang, W. A high-order imaging algorithm for high-resolution spaceborne SAR based on a modified equivalent squint range model. IEEE Trans. Geosci. Remote Sens. 2014, 53, 1225–1235. [Google Scholar] [CrossRef] [Green Version]
- Moreira, A.; Prats-Iraola, P.; Younis, M.; Krieger, G.; Hajnsek, I.; Papathanassiou, K.P. A tutorial on synthetic aperture radar. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–43. [Google Scholar] [CrossRef] [Green Version]
- Sun, Y.; Liu, Z.; Todorovic, S.; Li, J. Adaptive boosting for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2007, 43, 112–125. [Google Scholar] [CrossRef]
- Zhao, Q.; Principe, J.C. Support vector machines for SAR automatic target recognition. IEEE Trans. Aerosp. Electron. Syst. 2001, 37, 643–654. [Google Scholar] [CrossRef] [Green Version]
- Yuan, X.; Tang, T.; Xiang, D.; Li, Y.; Su, Y. Target recognition in SAR imagery based on local gradient ratio pattern. Int. J. Remote Sens. 2014, 35, 857–870. [Google Scholar] [CrossRef]
- Chen, S.; Wang, H. SAR target recognition based on deep learning. In Proceedings of the 2014 International Conference on Data Science and Advanced Analytics (DSAA), Shanghai, China, 30 October–2 November 2014; pp. 541–547. [Google Scholar]
- Tian, Z.; Zhan, R.; Hu, J.; Zhang, J. SAR ATR Based on Convolutional Neural Network. J. Radars 2016, 5, 4806–4817. [Google Scholar]
- Chen, S.; Wang, H.; Xu, F.; Jin, Y.-Q. Target Classification Using the Deep Convolutional Networks for SAR Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 857–870. [Google Scholar] [CrossRef]
- El-Darymli, K.; Gill, E.; Mcguire, P.; Power, D.; Moloney, C. Automatic Target Recognition in Synthetic Aperture Radar Imagery: A State-of-the-Art Review. IEEE Access 2016, 4, 6014–6058. [Google Scholar] [CrossRef] [Green Version]
- Bhanu, B.; Jones, G. Exploiting azimuthal variance of scatterers for multiple-look SAR recognition. Proc. SPIE- Int. Soc. Opt. Eng. 2002, 4727, 290–298. [Google Scholar]
- Keydel, E.; Lee, S.; Moore, J. MSTAR extended operating conditions: A tutorial. In Algorithms for Synthetic Aperture Radar Imagery III; International Society for Optics and Photonics: Washington, DC, USA, 1996; Volume 2757, pp. 228–242. [Google Scholar]
- Brendel, G.; Horowitz, L. Benefits of aspect diversity for SAR ATR: Fundamental and experimental results. Proc. SPIE- Int. Soc. Opt. Eng. 2000, 4053, 567–578. [Google Scholar]
- Ding, B.; Wen, G. Exploiting multi-view SAR images for robust target recognition. Remote Sens. 2017, 9, 1150. [Google Scholar] [CrossRef] [Green Version]
- Walterscheid, I.; Brenner, A. Multistatic and multi-aspect SAR data acquisition to improve image interpretation. IEEE Int. Geosci. Remote Sens. Symp. 2013, 4194–4197. [Google Scholar]
- Pei, J.; Huang, Y.; Huo, W.; Zhang, Y.; Yang, J.; Yeo, T.-S. SAR Automatic Target Recognition Based on Multiview Deep Learning Framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 2196–2210. [Google Scholar] [CrossRef]
- Zhao, P.; Liu, K.; Zou, H.; Zhen, X. Multi-Stream Convolutional Neural Network for SAR Automatic Target Recognition. Remote Sens. 2018, 10, 1473. [Google Scholar] [CrossRef] [Green Version]
- Zhang, F.; Fu, Z.; Zhou, Y.; Hu, W.; Hong, W. Multi-aspect SAR target recognition based on space-fixed and space-varying scattering feature joint learning. Remote Sens. Lett. 2019, 10, 998–1007. [Google Scholar] [CrossRef]
- Wang, B.; Pan, Z.; Hu, Y.; Ma, W. SAR Target Recognition Based on Siamese CNN with Small Scale Dataset. Radar Sci. Techol. 2019, 17, 603–609+615. [Google Scholar]
- Zhang, F.; Wang, Y.; Ni, J.; Zhou, Y.; Hu, W. SAR Target Small Sample Recognition Based on CNN Cascaded Features and AdaBoost Rotation Forest. IEEE Geosci. Remote Sens. Lett. 2020, 17, 1008–1012. [Google Scholar] [CrossRef]
- Zhai, Y.; Deng, W.; Lan, T.; Sun, B.; Ying, Z.; Gan, J. MFFA-SARNET: Deep Transferred Multi-Level Feature Fusion Attention Network with Dual Optimized Loss for Small-Sample SAR ATR. Remote Sens. 2020, 12, 1385. [Google Scholar] [CrossRef]
- Snell, J.; Swersky, K.; Zemel, R. Prototypical networks for few-shot learning. arxiv 2017, arXiv:1703.05175. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. Int. Conf. Mach. Learn. PMLR 2019, 6105–6114. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residu·al learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1770–1778. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arxiv 2014, arXiv:1409.1556. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
- Wang, Q.; Wu, B.; Zhu, P. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–18 June 2020. [Google Scholar]
- Dai, J.; Qi, H.; Xiong, Y. Deformable convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
- Xu, H.; Gu, H. A new algorithm of SAR target recognition based on advance deep learning neural network. Comput. Model. New Technol. 2014, 18, 25–30. [Google Scholar]
- Müller, R.; Kornblith, S.; Hinton, G. When does label smoothing help? arxiv 2019, arXiv:1906.02629. [Google Scholar]
- Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 499–515. [Google Scholar]
- Laurens, V.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
- Zhang, Z.; Wang, H.; Xu, F.; Jin, Y.-Q. Complex-Valued Convolutional Neural Network and Its Application in Polarimetric SAR Image Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 7177–7188. [Google Scholar] [CrossRef]
Stage | Operator | Channels | Layers |
---|---|---|---|
1 | Conv3 × 3 | 32 | 1 |
2 | MBConv1 | 16 | 1 |
3 | MBConv6 | 24 | 2 |
4 | MBConv6 | 40 | 2 |
5 | MBConv6 | 80 | 3 |
6 | MBConv6 | 112 | 3 |
7 | MBConv6 | 192 | 4 |
8 | MBConv6 | 320 | 1 |
9 | Conv1d&Pooling&FC | k | 1 |
Target | Train Dataset Size | Test Dataset Size |
---|---|---|
2S1 | 1162 | 1034 |
BMP2 | 883 | 634 |
BRDM_2 | 1158 | 1040 |
BTR70 | 889 | 649 |
BTR60 | 978 | 667 |
D7 | 1162 | 1037 |
T62 | 1162 | 1032 |
T72 | 874 | 642 |
ZIL131 | 1162 | 1034 |
ZSU_234 | 1162 | 1040 |
Total | 10,592 | 8809 |
Target | Train Dataset Size | Test Dataset Size |
---|---|---|
2S1 | 1166 | 1122 |
BRDM_2 | 1162 | 1118 |
T72 | 913 | 1122 |
ZSU_234 | 1166 | 1122 |
Total | 4407 | 4484 |
Type | Parameter |
---|---|
Batch Size | 32 |
Optimizer | Adam |
Adam’s Learning Rate | 0.001 |
Center Loss’ Optimizer | SGD |
SGD’s Learning Rate | 0.5 |
Center Loss’ Hyper-Parameter | 0.01 |
Cosine Decay Max Epoch | 100 |
Epoch | 100 |
Dataset | Direct Training Acc. | Our Method Acc. | Increase in Acc. |
---|---|---|---|
SOC-150 | 77.06% | 99.84% | 21.34% |
SOC-250 | 88.65% | 99.97% | 11.32% |
Complete SOC Dataset | 100% | - | - |
Dataset | Direct Training Acc. | Our Method Acc. | Increase in Acc. |
---|---|---|---|
EOC1-100 | 68.58% | 98.38% | 29.80% |
EOC1-200 | 84.30% | 99.26% | 14.96% |
Complete EOC1 Dataset | 99.45% | - | - |
Dataset | Direct Training Acc. | Our Method Acc. | Increase in Acc. |
---|---|---|---|
EOC23-100 | 92.56% | 99.39% | 6.83% |
EOC23-200 | 96.47% | 99.67% | 3.20% |
Complete EOC23 Dataset | 99.96% | - | - |
Dataset | Direct Training Acc. | Our Method Acc. | Increase in Acc. |
---|---|---|---|
EOC23-100 | 99.03% | 99.28% | 0.25% |
EOC23-200 | 99.54% | 99.69% | 0.15% |
Complete EOC23 Dataset | 99.78% | - | - |
SOC-150 Dataset | SOC-250 Dataset | |||
---|---|---|---|---|
Index | Accuracy | Increase in Accuracy | Accuracy | Increase in Accuracy |
1 | 77.06% | - | 88.65% | -% |
2 | 86.30% | 9.24% | 95.45% | 6.80% |
3 | 90.62% | 4.32% | 96.34% | 0.89% |
4 | 94.74% | 4.12% | 98.49% | 2.15% |
5 | 97.60% | 2.86% | 99.42% | 0.93% |
6 | 97.91% | 0.31% | 99.56% | 0.14% |
7 | 98.40% | 0.49% | 99.97% | 0.41% |
Index | Time Costs | Increase in Time Costs |
---|---|---|
1 | 0.774 ms | - |
2 | 1.441 ms | 0.667 ms |
3 | 1.519 ms | 0.078 ms |
4 | 1.602 ms | 0.083 ms |
5 | 1.702 ms | 0.100 ms |
6 | 1.739 ms | 0.037 ms |
7 | 2.968 ms | 1.229 ms |
Model | Dataset | Direct Training Acc. | Our Method Acc. | Increase in Acc. |
---|---|---|---|---|
ResNet18 | SOC-150 | 87.10% | 98.25% | 11.15% |
SOC-250 | 92.85% | 99.95% | 7.10% | |
VggNet11 | SOC-150 | 80.87% | 99.13% | 18.26% |
SOC-250 | 90.35% | 99.80% | 9.45% | |
AlexNet | SOC-150 | 85.52% | 96.17% | 10.65% |
SOC-250 | 94.15% | 98.20% | 4.05% |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhao, P.; Huang, L.; Xin, Y.; Guo, J.; Pan, Z. Multi-Aspect SAR Target Recognition Based on Prototypical Network with a Small Number of Training Samples. Sensors 2021, 21, 4333. https://doi.org/10.3390/s21134333
Zhao P, Huang L, Xin Y, Guo J, Pan Z. Multi-Aspect SAR Target Recognition Based on Prototypical Network with a Small Number of Training Samples. Sensors. 2021; 21(13):4333. https://doi.org/10.3390/s21134333
Chicago/Turabian StyleZhao, Pengfei, Lijia Huang, Yu Xin, Jiayi Guo, and Zongxu Pan. 2021. "Multi-Aspect SAR Target Recognition Based on Prototypical Network with a Small Number of Training Samples" Sensors 21, no. 13: 4333. https://doi.org/10.3390/s21134333
APA StyleZhao, P., Huang, L., Xin, Y., Guo, J., & Pan, Z. (2021). Multi-Aspect SAR Target Recognition Based on Prototypical Network with a Small Number of Training Samples. Sensors, 21(13), 4333. https://doi.org/10.3390/s21134333