Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,199)

Search Parameters:
Keywords = radar network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 1850 KiB  
Review
Deep Learning for Automatic Detection of Volcanic and Earthquake-Related InSAR Deformation
by Xu Liu, Yingfeng Zhang, Xinjian Shan, Zhenjie Wang, Wenyu Gong and Guohong Zhang
Remote Sens. 2025, 17(4), 686; https://doi.org/10.3390/rs17040686 - 18 Feb 2025
Viewed by 94
Abstract
Interferometric synthetic aperture radar (InSAR) technology plays a crucial role in monitoring surface deformation and has become widely used in volcanic and earthquake research. With the rapid advancement of satellite technology, InSAR now generates vast volumes of deformation data. Deep learning has revolutionized [...] Read more.
Interferometric synthetic aperture radar (InSAR) technology plays a crucial role in monitoring surface deformation and has become widely used in volcanic and earthquake research. With the rapid advancement of satellite technology, InSAR now generates vast volumes of deformation data. Deep learning has revolutionized data analysis, offering exceptional capabilities for processing large datasets. Leveraging these advancements, automatic detection of volcanic and earthquake deformation from extensive InSAR datasets has emerged as a major research focus. In this paper, we first introduce several representative deep learning architectures commonly used in InSAR data analysis, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), and Transformer networks. Each architecture offers unique advantages for addressing the challenges of InSAR data. We then systematically review recent progress in the automatic detection and identification of volcanic and earthquake deformation signals from InSAR images using deep learning techniques. This review highlights two key aspects: the design of network architectures and the methodologies for constructing datasets. Finally, we discuss the challenges in automatic detection and propose potential solutions. This study aims to provide a comprehensive overview of the current applications of deep learning for extracting InSAR deformation features, with a particular focus on earthquake and volcanic monitoring. Full article
Show Figures

Figure 1

Figure 1
<p>InSAR data processing based on deep learning. (<b>a</b>) The primary deep learning architectures utilized in InSAR data processing, including CNNs, RNNs, GANs, and Transformers. (<b>b</b>) DL is applied to various stages of InSAR data processing, including deformation detection, atmospheric correction, phase filtering, and phase unwrapping.</p>
Full article ">Figure 2
<p>The main architectures of CNNs, RNNs, GANs, and Transformer networks. (<b>a</b>) CNNs primarily consist of an input layer, convolutional layers, pooling layers, fully connected layers, and an output layer. (<b>b</b>) RNNs consist of input layers, recurrent hidden layers, and an output layer for sequence tasks. (<b>c</b>) GANs consist of a generator and a discriminator, which are trained together in a competitive manner. (<b>d</b>) Transformers consist of an encoder and a decoder, both using self-attention and feed-forward layers.</p>
Full article ">Figure 3
<p>Different learning processes between traditional machine learning and transfer learning. (<b>a</b>) Traditional machine learning approaches learn each task independently, starting from scratch. (<b>b</b>) Transfer learning utilizes knowledge gained from previous tasks and applies it to a target task.</p>
Full article ">Figure 4
<p>Data augmentation methods. (<b>a</b>) Geometric transformation-based data augmentation involves techniques like zoom, rotation, mirroring, and flipping to expand the training datasets. (<b>b</b>) Pixel-level transformation-based data augmentation modifies individual pixel values, such as brightness, contrast, and color, to enhance the datasets. (<b>c</b>) Filtering-based data augmentation involves applying filters like blurring, sharpening, and noise to diversify the training datasets. The original InSAR interferogram data were downloaded from the COMET-LiCS Sentinel-1 InSAR portal (<a href="https://comet.nerc.ac.uk/comet-lics-portal/" target="_blank">https://comet.nerc.ac.uk/comet-lics-portal/</a> (accessed on 1 December 2024)).</p>
Full article ">
25 pages, 4721 KiB  
Article
Human Respiration and Motion Detection Based on Deep Learning and Signal Processing Techniques to Support Search and Rescue Teams
by Özden Niyaz, Mehmet Ziya Erenoğlu, Ahmet Serdar Türk, Sultan Aldirmaz Colak, Burcu Erkmen and Nurhan Türker Tokan
Appl. Sci. 2025, 15(4), 2097; https://doi.org/10.3390/app15042097 - 17 Feb 2025
Viewed by 154
Abstract
The quick and effective detection of humans trapped under debris is crucial in search and rescue operations. This study explores the use of antennas operating within the 150–650 MHz frequency range to identify human respiration and movement under building wreckage. A debris model [...] Read more.
The quick and effective detection of humans trapped under debris is crucial in search and rescue operations. This study explores the use of antennas operating within the 150–650 MHz frequency range to identify human respiration and movement under building wreckage. A debris model consisting of construction materials was generated at the laboratory, and attenuation characteristics were observed to set ideal operating frequencies. Time-dependent transmission coefficient data were collected over 20 s and processed using short-time Fourier transform, wavelet transform, and empirical mode decomposition for time-frequency analysis. To enhance signal clarity, denoising techniques were applied before the radar signals were categorized into three classes: empty debris, human respiration, and human movement. Generative adversarial networks augmented environmental noise data to enrich training datasets comprising nine subsets. Deep learning models, including temporal convolutional networks, long short-term memory, and convolutional neural networks, were employed for classification. Hyperparameter optimization via random search further refined model performance. Results indicate that the convolutional neural networks using short-time Fourier transform data consistently achieved the highest classification accuracy across subsets. These findings demonstrate the potential of combining radar with deep learning for reliable human detection under debris, advancing rescue efforts in disaster scenarios. Full article
Show Figures

Figure 1

Figure 1
<p>Radar based human detection system.</p>
Full article ">Figure 2
<p>Physically constructed debris model in our laboratory. (<b>a</b>) Front view. (<b>b</b>) Side view.</p>
Full article ">Figure 3
<p>Variation of the raw measured magnitude and phase of <math display="inline"><semantics> <msub> <mi>S</mi> <mn>21</mn> </msub> </semantics></math> signals at 405 MHz for different scenarios (<b>a</b>) Empty debris. (<b>b</b>) Respiration. (<b>c</b>) Movement. (<b>d</b>) Combined plots.</p>
Full article ">Figure 4
<p>Analysis of the <math display="inline"><semantics> <msub> <mi>S</mi> <mn>21</mn> </msub> </semantics></math> signals measured for different orientation angles of the subject: (<b>a</b>) FT of the measured phase of <math display="inline"><semantics> <msub> <mi>S</mi> <mn>21</mn> </msub> </semantics></math> signals at 405 MHz. (<b>b</b>) Average standard deviation of the phase and magnitude at different frequencies.</p>
Full article ">Figure 5
<p>STFT images of the measured signal for 20 s: (<b>a</b>) when the debris is empty. A subject is inside the debris with orientation of: (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>0</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>45</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>90</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>e</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>180</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>f</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>270</mn> <mo>°</mo> </msup> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>CWT images of the measured signal for 20 s: (<b>a</b>) when the debris is empty. A subject is inside the debris with orientation of: (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>0</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>c</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>45</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>d</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>90</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>e</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>180</mn> <mo>°</mo> </msup> </mrow> </semantics></math>. (<b>f</b>) <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>270</mn> <mo>°</mo> </msup> </mrow> </semantics></math>.</p>
Full article ">Figure 7
<p>Selected IMFs of the EMD analysis for a measured signal of a moving subject inside the debris with <math display="inline"><semantics> <mrow> <mi>ϕ</mi> <mo>=</mo> <msup> <mn>0</mn> <mo>°</mo> </msup> </mrow> </semantics></math> orientation angle.</p>
Full article ">Figure 8
<p>Comparison of original and GAN-augmented noise signals (Magnitude).</p>
Full article ">Figure 9
<p>Comparison of original and GAN-augmented noise signals (Phase).</p>
Full article ">Figure 10
<p>LSTM of original and GAN−: (<b>a</b>) Network. (<b>b</b>) Memory block with one cell.</p>
Full article ">Figure 11
<p>Convolutional neural network topology.</p>
Full article ">Figure 12
<p>Confusion matrices of the DL models for Subset 3: (<b>a</b>) LSTM. (<b>b</b>) TCN.</p>
Full article ">Figure 13
<p>Confusion matrices of the DL models for Subset 4: (<b>a</b>) LSTM. (<b>b</b>) TCN. (<b>c</b>) CNN (with STFT data). (<b>d</b>) CNN (with CWT data).</p>
Full article ">
28 pages, 8850 KiB  
Article
Real-Time Runway Detection Using Dual-Modal Fusion of Visible and Infrared Data
by Lichun Yang, Jianghao Wu, Hongguang Li, Chunlei Liu and Shize Wei
Remote Sens. 2025, 17(4), 669; https://doi.org/10.3390/rs17040669 - 16 Feb 2025
Viewed by 158
Abstract
Advancements in aviation technology have made intelligent navigation systems essential for improving flight safety and efficiency, particularly in low-visibility conditions. Radar and GPS systems face limitations in bad weather, making visible–infrared sensor fusion a promising alternative. This study proposes a salient object detection [...] Read more.
Advancements in aviation technology have made intelligent navigation systems essential for improving flight safety and efficiency, particularly in low-visibility conditions. Radar and GPS systems face limitations in bad weather, making visible–infrared sensor fusion a promising alternative. This study proposes a salient object detection (SOD) method that integrates visible and infrared sensors for robust airport runway detection in complex environments. We introduce a large-scale visible–infrared runway dataset (RDD5000) and develop a SOD algorithm capable of detecting salient targets from unaligned visible and infrared images. To enable real-time processing, we design a lightweight dual-modal fusion network (DCFNet) with an independent–shared encoder and a cross-layer attention mechanism to enhance feature extraction and fusion. Experimental results show that the MobileNetV2-based lightweight version achieves 155 FPS on a single GPU, significantly outperforming previous methods such as DCNet (4.878 FPS) and SACNet (27 FPS), making it suitable for real-time deployment on airborne systems. This work offers a novel and efficient solution for intelligent navigation in aviation. Full article
Show Figures

Figure 1

Figure 1
<p>The examples of several airport runway datasets: (<b>a</b>) airport runway datasets using remote sensing imagery [<a href="#B44-remotesensing-17-00669" class="html-bibr">44</a>,<a href="#B45-remotesensing-17-00669" class="html-bibr">45</a>,<a href="#B46-remotesensing-17-00669" class="html-bibr">46</a>], (<b>b</b>) airport runway datasets generated by simulation [<a href="#B47-remotesensing-17-00669" class="html-bibr">47</a>,<a href="#B48-remotesensing-17-00669" class="html-bibr">48</a>].</p>
Full article ">Figure 2
<p>The examples of the proposed RDD5000 dataset consisting of 10 sets of visible images (first row and fifth row), visible ground truth (second row and sixth row), infrared images (third row and seventh row), and infrared ground truth (fourth row and eighth row).</p>
Full article ">Figure 3
<p>The overall architecture of the proposed DCFNet.</p>
Full article ">Figure 4
<p>The proposed MAFM framework.</p>
Full article ">Figure 5
<p>Examples of haze image generation.</p>
Full article ">Figure 6
<p>Visual comparison of speed and accuracy on the RDD5000 datasets.</p>
Full article ">Figure 7
<p>PR curves and threshold F-measure curves (from <b>left</b> to <b>right</b>) of different models on the RDD5000 visible dataset.</p>
Full article ">Figure 8
<p>Visual comparisons with other SOTA methods under different challenging scenarios, including small objects (Rows 1 and 2), medium objects (Rows 3 and 4), and large objects (Rows 5 and 6).</p>
Full article ">Figure 9
<p>Max F-measure scores and MAE on RDD5000 in the whole training procedure to verify the effectiveness of the MFEM.</p>
Full article ">Figure 10
<p>Visual comparison of the feature heatmaps with/without the MFEM.</p>
Full article ">Figure 11
<p>Visual comparison of the features with/without AM in the MAFM.</p>
Full article ">
12 pages, 839 KiB  
Article
ISAR Image Quality Assessment Based on Visual Attention Model
by Jun Zhang, Zhicheng Zhao and Xilan Tian
Appl. Sci. 2025, 15(4), 1996; https://doi.org/10.3390/app15041996 - 14 Feb 2025
Viewed by 189
Abstract
The quality of ISAR (Inverse Synthetic Aperture Radar) images has a significant impact on the detection and recognition of targets. Therefore, ISAR image quality assessment is a fundamental prerequisite and primary link in the utilization of ISAR images. Previous ISAR image quality assessment [...] Read more.
The quality of ISAR (Inverse Synthetic Aperture Radar) images has a significant impact on the detection and recognition of targets. Therefore, ISAR image quality assessment is a fundamental prerequisite and primary link in the utilization of ISAR images. Previous ISAR image quality assessment methods typically extract hand-crafted features or use simple multi-layer networks to extract local features. Hand-crafted features and local features from networks usually lack the global information of ISAR images. Furthermore, most deep neural networks obtain feature representations by abridging the prediction quality score and the ground truth, neglecting to explore the strong correlations between features and quality scores in the stage of feature extraction. This study proposes a Gramin Transformer to explore the similarity and diversity of features extracted from different images, thus obtaining features containing quality-related information. The Gramin matrix of features is computed to obtain the score token through the self-attention layer. It prompts the network to learn more discriminative features, which are closely associated with quality scores. Despite the Transformer architecture’s ability to extract global information, the Channel Attention Block (CAB) can capture complementary information from different channels in an image, aggregating and mining information from these channels to provide a more comprehensive evaluation of ISAR images. ISAR images are formed from target scattering points with a background containing substantial silent noise, and the Inter-Region Attention Block (IRAB) is utilized to extract local scattering point features, which decide the clarity of target. In addition, extensive experiments are conducted on the ISAR image dataset (including space stations, ships, aircraft, etc.). The evaluation results of our method on the dataset are significantly superior to those of traditional feature extraction methods and existing image quality assessment methods. Full article
Show Figures

Figure 1

Figure 1
<p>The architecture of the proposed model. The image is partitioned into 8 × 8-sized patches. Then, the linear projection layer performs a convolution operation on these patches to acquire patch embeddings, which can be processed by the Transformer encoder. The Gram–T model computes the Gramin matrix of extracted features to obtain score tokens through one attention layer. Moreover, the CAB and IRAB strength interactions between channels and regions within features are output by the Transformer encoder. Finally, the prediction score from the IRAB is added by a score token from Gram–T to obtain the final score.</p>
Full article ">Figure 2
<p>CAB module.</p>
Full article ">Figure 3
<p>IRAB module.</p>
Full article ">Figure 4
<p>Display of the results in the training process.</p>
Full article ">Figure 5
<p>Example images from ISAR image dataset. The first caption row below the images refers to the ground truth scores of the according images, and the second row refers to prediction scores.</p>
Full article ">Figure 6
<p>Attention heatmap analysis. The attention heatmap is on the left, and the original image is on the right. In our experiments, the blue color in the attention heatmap represents the high weight, and the red color represents low weight.</p>
Full article ">
22 pages, 2839 KiB  
Article
Narrowband Radar Micromotion Targets Recognition Strategy Based on Graph Fusion Network Constructed by Cross-Modal Attention Mechanism
by Yuanjie Zhang, Ting Gao, Hongtu Xie, Haozong Liu, Mengfan Ge, Bin Xu, Nannan Zhu and Zheng Lu
Remote Sens. 2025, 17(4), 641; https://doi.org/10.3390/rs17040641 - 13 Feb 2025
Viewed by 283
Abstract
In the domain of micromotion target recognition, target characteristics can be extracted through either broadband or narrowband radar echoes. However, due to technical limitations and cost constraints in acquiring broadband radar waveform data, recognition can often only be performed through narrowband radar waveforms. [...] Read more.
In the domain of micromotion target recognition, target characteristics can be extracted through either broadband or narrowband radar echoes. However, due to technical limitations and cost constraints in acquiring broadband radar waveform data, recognition can often only be performed through narrowband radar waveforms. To fully utilize the information embedded within narrowband radar waveforms, it is necessary to conduct in-depth research on multi-dimensional features of micromotion targets, including radar cross-sections (RCSs), time frequency (TF) images, and cadence velocity diagrams (CVDs). To address the limitations of existing identification methodologies in achieving accurate recognition with narrowband echoes, this paper proposes a graph fusion network based on a cross-modal attention mechanism, termed GF-AM Net. The network first adopts convolutional neural networks (CNNs) to extract unimodal features from RCSs, TF images, and CVDs independently. Subsequently, a cross-modal attention mechanism integrates these extracted features into a graph structure, achieving multi-level interactions among unimodal, bimodal, and trimodal features. Finally, the fused features are input into a classification module to accomplish narrowband radar micromotion target identification. Experimental results demonstrate that the proposed methodology successfully captures potential correlations between modal features by incorporating cross-modal multi-level information interactions across different processing stages, exhibiting exceptional accuracy and robustness in narrowband radar micromotion target identification tasks. Full article
(This article belongs to the Special Issue Ocean Remote Sensing Based on Radar, Sonar and Optical Techniques)
Show Figures

Figure 1

Figure 1
<p>Diagram of the micromotion model.</p>
Full article ">Figure 2
<p>Process of converting the narrowband radar echoes into the TF image and CVD.</p>
Full article ">Figure 3
<p>Overall structure of the GF-AM Net.</p>
Full article ">Figure 4
<p>Structure diagram of the unimodal feature extraction network.</p>
Full article ">Figure 5
<p>Cross-modal attention mechanism in the graph fusion module.</p>
Full article ">Figure 6
<p>3D structure of the four types of micromotion targets. From left to right, they are Target 1, Target 2, Target 3, and Target 4.</p>
Full article ">Figure 7
<p>Static narrowband radar echo data for four types of micromotion targets. (<b>a</b>) Target 1; (<b>b</b>) Target 2; (<b>c</b>) Target 3; (<b>d</b>) Target 4.</p>
Full article ">Figure 8
<p>RCS for four types of micromotion targets. (<b>a</b>) Target 1; (<b>b</b>) Target 2; (<b>c</b>) Target 3; (<b>d</b>) Target 4.</p>
Full article ">Figure 9
<p>TF image for four types of micromotion targets. (<b>a</b>) Target 1; (<b>b</b>) Target 2; (<b>c</b>) Target 3; (<b>d</b>) Target 4.</p>
Full article ">Figure 10
<p>CVD for four types of micromotion targets. (<b>a</b>) Target 1; (<b>b</b>) Target 2; (<b>c</b>) Target 3; (<b>d</b>) Target 4.</p>
Full article ">Figure 11
<p>Comparison of training curves for different methods.</p>
Full article ">Figure 12
<p>Confusion matrix of different methods. (<b>a</b>) RCS Net; (<b>b</b>) Mobilenetvit_1; (<b>c</b>) Mobilenetvit_2; (<b>d</b>) TLPS Net; (<b>e</b>) ACM-FR Net; (<b>f</b>) GF-AM Net.</p>
Full article ">Figure 13
<p>Visualization of the corresponding features extracted by the different networks. (<b>a</b>) RCS Net; (<b>b</b>) Mobilenetvit_1; (<b>c</b>) Mobilenetvit_2; (<b>d</b>) TLPS Net; (<b>e</b>) ACM-FR Net; (<b>f</b>) GF-AM Net.</p>
Full article ">Figure 14
<p>Training curve comparison in ablation experiments.</p>
Full article ">
17 pages, 3052 KiB  
Article
Estimation of Daylily Leaf Area Index by Synergy Multispectral and Radar Remote-Sensing Data Based on Machine-Learning Algorithm
by Minhuan Hu, Jingshu Wang, Peng Yang, Ping Li, Peng He and Rutian Bi
Agronomy 2025, 15(2), 456; https://doi.org/10.3390/agronomy15020456 - 13 Feb 2025
Viewed by 276
Abstract
Rapid and accurate leaf area index (LAI) determination is important for monitoring daylily growth, yield estimation, and field management. Because of low estimation accuracy of empirical models based on single-source data, we proposed a machine-learning algorithm combining optical and microwave remote-sensing data as [...] Read more.
Rapid and accurate leaf area index (LAI) determination is important for monitoring daylily growth, yield estimation, and field management. Because of low estimation accuracy of empirical models based on single-source data, we proposed a machine-learning algorithm combining optical and microwave remote-sensing data as well as the random forest regression (RFR) importance score to select features. A high-precision LAI estimation model for daylilies was constructed by optimizing feature combinations. The RFR importance score screened the top five important features, including vegetation indices land surface water index (LSWI), generalized difference vegetation index (GDVI), normalized difference yellowness index (NDYI), and backscatter coefficients VV and VH. Vegetation index features characterized canopy moisture and the color of daylilies, and the backscatter coefficient reflected dielectric properties and geometric structure. The selected features were sensitive to daylily LAI. The RFR algorithm had good anti-noise performance and strong fitting ability; thus, its accuracy was better than the partial least squares regression and artificial neural network models. Synergistic optical and microwave data more comprehensively reflected the physical and chemical properties of daylilies, making the RFR-VI-BC05 model after feature selection better than the others ( r = 0.711, RMSE = 0.498, and NRMSE = 9.10%). This study expanded methods for estimating daylily LAI by combining optical and radar data, providing technical support for daylily management. Full article
Show Figures

Figure 1

Figure 1
<p>Location and sampling distribution of the study area.</p>
Full article ">Figure 2
<p>Technical route.</p>
Full article ">Figure 3
<p>The importance score of features. (<b>a</b>) shows the importance scores of the vegetation index, (<b>b</b>) displays the importance scores of the backscattering coefficient, and (<b>c</b>) presents the combined importance scores of both the vegetation index and backscattering coefficient.</p>
Full article ">Figure 4
<p>Regression prediction models based on radar data.</p>
Full article ">Figure 5
<p>Regression prediction models based on optical data.</p>
Full article ">Figure 6
<p>Regression prediction models based on multisource remote-sensing data.</p>
Full article ">Figure 7
<p>LAI inversion results of daylily and classification of LAI in each township.</p>
Full article ">
18 pages, 8926 KiB  
Article
Research on Damage Detection Methods for Concrete Beams Based on Ground Penetrating Radar and Convolutional Neural Networks
by Ning Liu, Ya Ge, Xin Bai, Zi Zhang, Yuhao Shangguan and Yan Li
Appl. Sci. 2025, 15(4), 1882; https://doi.org/10.3390/app15041882 - 12 Feb 2025
Viewed by 359
Abstract
Ground penetrating radar (GPR) is a mature and important research method in the field of structural non-destructive testing. However, when the detection target scale is small and the amount of data collected is limited, it poses a serious challenge for this research method. [...] Read more.
Ground penetrating radar (GPR) is a mature and important research method in the field of structural non-destructive testing. However, when the detection target scale is small and the amount of data collected is limited, it poses a serious challenge for this research method. In order to verify the applicability of typical one-dimensional radar signals combined with convolutional neural networks (CNN) in the non-destructive testing of concrete structures, this study created concrete specimens with embedded defects (voids, non-dense solids, and cracks) commonly found in concrete structures in a laboratory setting. High-frequency GPR equipment is used for data acquisition, A-scan data corresponding to different defects is extracted as a training set, and appropriate labeling is carried out. The extracted original radar signals were taken as the input of the CNN model. At the same time, in order to improve the sensitivity of the CNN models to specific damage types, the spectrums of A-scan are also used as part of the training datasets of the CNN models. In this paper, two CNN models with different dimensions are used to train the datasets and evaluate the classification results; one is the traditional one-dimensional CNN model, and the other is the classical two-dimensional CNN architecture AlexNet. In addition, the finite difference time domain (FDTD) model of three-dimensional complex media is established by gprMax, and the propagation characteristics of GPR in concrete media are simulated. The results of applying this method to both simulated and experimental data show that combining the A-scan data of ground penetrating radar and their spectrums as input with the CNN model can effectively identify different types of damage and defects inside the concrete structure. Compared with the one-dimensional CNN model, AlexNet has obvious advantages in extracting complex signal features and processing high-dimensional data. The feasibility of this method in the research field of damage detection of concrete structures has been verified. Full article
(This article belongs to the Special Issue Ground Penetrating Radar: Data, Imaging, and Signal Analysis)
Show Figures

Figure 1

Figure 1
<p>Diagram of the GPR detection concrete distress principle.</p>
Full article ">Figure 2
<p>Process of the one-dimensional convolution layer.</p>
Full article ">Figure 3
<p>Process of the two-dimensional convolution layer.</p>
Full article ">Figure 4
<p>Process of the pooling layer.</p>
Full article ">Figure 5
<p>The diagram is one-dimensional.</p>
Full article ">Figure 6
<p>The diagram is two-dimensional.</p>
Full article ">Figure 7
<p>(<b>a</b>) H_B0: reinforced concrete beam; (<b>b</b>) H_BF1: 40 mm diameter PVC pipe and 60 mm side length of non-confined solid; (<b>c</b>) H_BF2: 25 mm diameter PVC pipe and 30 mm side length plastic foam.</p>
Full article ">Figure 8
<p>A heterogeneous numerical model of concrete.</p>
Full article ">Figure 9
<p>Models of concrete beams with defects: (<b>a</b>) H_B0; (<b>b</b>) H_BF1; (<b>c</b>) H_BF2.</p>
Full article ">Figure 10
<p>Concrete surface line tracks.</p>
Full article ">Figure 11
<p>The setting of prefabricated defects in concrete beams: (<b>a</b>) non-dense material, void, and (<b>b</b>) cracks generated during the experiment.</p>
Full article ">Figure 12
<p>(<b>a</b>) GSSI GPR equipment and (<b>b</b>) measurement process.</p>
Full article ">Figure 13
<p>Radar profile features of different defects (Figure (<b>a</b>–<b>c</b>) are simulation results, while Figure (<b>d</b>–<b>f</b>) are experiment results).</p>
Full article ">Figure 14
<p>Comparison of (<b>a</b>) simulated A-scan and (<b>b</b>) spectrum of simulated A-scan for three different defect types; (<b>c</b>) experimental A-scan and (<b>d</b>) spectrum of experimental A-scan for three different defect types.</p>
Full article ">Figure 15
<p>The fitting curves of the training process of two kinds of convolutional networks for (<b>a</b>) the one-dimensional CNN and (<b>b</b>) the two-dimensional CNN.</p>
Full article ">Figure 16
<p>Classification results of one-dimensional CNN model for simulated data for (<b>a</b>) the A-scan and (<b>b</b>) the spectrum of the A-scan.</p>
Full article ">Figure 17
<p>Classification results of two-dimensional CNN model for simulated data for (<b>a</b>) the A-scan and (<b>b</b>) the spectrum of the A-scan.</p>
Full article ">Figure 18
<p>Classification results of one-dimensional CNN model for experimental data for (<b>a</b>) the A-scan and (<b>b</b>) the spectrum of the A-scan.</p>
Full article ">Figure 19
<p>Classification results of two-dimensional CNN model for experimental data for (<b>a</b>) the A-scan and (<b>b</b>) the spectrum of the A-scan.</p>
Full article ">Figure 20
<p>Classification results of two-dimensional CNN model for merged data for (<b>a</b>) the A-scan and (<b>b</b>) the spectrum of the A-scan.</p>
Full article ">Figure 21
<p>Accuracy values for training on experimental and simulated data using (<b>a</b>) the one-dimensional CNN model and (<b>b</b>) the two-dimensional CNN model.</p>
Full article ">
23 pages, 5392 KiB  
Article
A Sliding Window-Based CNN-BiGRU Approach for Human Skeletal Pose Estimation Using mmWave Radar
by Yuquan Luo, Yuqiang He, Yaxin Li, Huaiqiang Liu, Jun Wang and Fei Gao
Sensors 2025, 25(4), 1070; https://doi.org/10.3390/s25041070 - 11 Feb 2025
Viewed by 308
Abstract
In this paper, we present a low-cost, low-power millimeter-wave (mmWave) skeletal joint localization system. High-quality point cloud data are generated using the self-developed BHYY_MMW6044 59–64 GHz mmWave radar device. A sliding window mechanism is introduced to extend the single-frame point cloud into multi-frame [...] Read more.
In this paper, we present a low-cost, low-power millimeter-wave (mmWave) skeletal joint localization system. High-quality point cloud data are generated using the self-developed BHYY_MMW6044 59–64 GHz mmWave radar device. A sliding window mechanism is introduced to extend the single-frame point cloud into multi-frame time-series data, enabling the full utilization of temporal information. This is combined with convolutional neural networks (CNNs) for spatial feature extraction and a bidirectional gated recurrent unit (BiGRU) for temporal modeling. The proposed spatio-temporal information fusion framework for multi-frame point cloud data fully exploits spatio-temporal features, effectively alleviates the sparsity issue of radar point clouds, and significantly enhances the accuracy and robustness of pose estimation. Experimental results demonstrate that the proposed system accurately detects 25 skeletal joints, particularly improving the positioning accuracy of fine joints, such as the wrist, thumb, and fingertip, highlighting its potential for widespread application in human–computer interaction, intelligent monitoring, and motion analysis. Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

Figure 1
<p>FMCW radar transmit and receive waveforms.</p>
Full article ">Figure 2
<p>FMCW radar RX antenna array and phase relationship.</p>
Full article ">Figure 3
<p>List and locations of 25 skeletal points.</p>
Full article ">Figure 4
<p>Overall flowchart of the human skeletal pose estimation system based on mmWave wave radar and CNN-BiGRU.</p>
Full article ">Figure 5
<p>Multi-frame point cloud temporal modeling based on sliding windows.</p>
Full article ">Figure 6
<p>Spatio-temporal information fusion network architecture based on CNN-BiGRU.</p>
Full article ">Figure 7
<p>mmWave radar structure.</p>
Full article ">Figure 8
<p>Experimental setup with one radar and one Kinect.</p>
Full article ">Figure 9
<p>Experimental environment.</p>
Full article ">Figure 10
<p>Average MAE for 25 human skeletal joints (<span class="html-italic">MARS</span> dataset).</p>
Full article ">Figure 11
<p>Average RMSE for 25 human skeletal joints (<span class="html-italic">MARS</span> dataset).</p>
Full article ">Figure 12
<p>Average MAE for 25 human skeletal joints (self-built dataset).</p>
Full article ">Figure 13
<p>Average RMSE for 25 human skeletal joints (self-built dataset).</p>
Full article ">Figure 14
<p>Demonstration of CNN-BiGRU reconstructing human skeletal joints from point cloud. From left to right, it shows radar point cloud, CNN-BiGRU prediction, and ground truth, respectively. The movements from top to bottom are left upper limb stretch, double upper limb stretch, left front lunge, right front lunge, and left lunge (self-built dataset).</p>
Full article ">Figure 15
<p>Average localization error for 25 human skeletal joints under different <math display="inline"><semantics> <mrow> <mi>s</mi> <mi>t</mi> <mi>e</mi> <mi>p</mi> </mrow> </semantics></math> value.</p>
Full article ">
24 pages, 13033 KiB  
Article
Detection of Parabolic Antennas in Satellite Inverse Synthetic Aperture Radar Images Using Component Prior and Improved-YOLOv8 Network in Terahertz Regime
by Liuxiao Yang, Hongqiang Wang, Yang Zeng, Wei Liu, Ruijun Wang and Bin Deng
Remote Sens. 2025, 17(4), 604; https://doi.org/10.3390/rs17040604 - 10 Feb 2025
Viewed by 330
Abstract
Inverse Synthetic Aperture Radar (ISAR) images of space targets and their key components are very important. However, this method suffers from numerous drawbacks, including a low Signal-to-Noise Ratio (SNR), blurred edges, significant variations in scattering intensity, and limited data availability, all of which [...] Read more.
Inverse Synthetic Aperture Radar (ISAR) images of space targets and their key components are very important. However, this method suffers from numerous drawbacks, including a low Signal-to-Noise Ratio (SNR), blurred edges, significant variations in scattering intensity, and limited data availability, all of which constrain its recognition capabilities. The terahertz (THz) regime has reflected excellent capacity for space detection in terms of showing the details of target structures. However, in ISAR images, as the observation aperture moves, the imaging features of the extended structures (ESs) undergo significant changes, posing challenges to the subsequent recognition performance. In this paper, a parabolic antenna is taken as the research object. An innovative approach for identifying this component is proposed by using the advantages of the Component Prior and Imaging Characteristics (CPICs) effectively. In order to tackle the challenges associated with component identification in satellite ISAR imagery, this study employs the Improved-YOLOv8 model, which was developed by incorporating the YOLOv8 algorithm, an adaptive detection head known as the Dynamic head (Dyhead) that utilizes an attention mechanism, and a regression box loss function called Wise Intersection over Union (WIoU), which addresses the issue of varying sample difficulty. After being trained on the simulated dataset, the model demonstrated a considerable enhancement in detection accuracy over the five base models, reaching an mAP50 of 0.935 and an mAP50-95 of 0.520. Compared with YOLOv8n, it improved by 0.192 and 0.076 in mAP50 and mAP50-95, respectively. Ultimately, the effectiveness of the suggested method is confirmed through the execution of comprehensive simulations and anechoic chamber tests. Full article
(This article belongs to the Special Issue Advanced Spaceborne SAR Processing Techniques for Target Detection)
Show Figures

Figure 1

Figure 1
<p>The overall framework diagram of the proposed method.</p>
Full article ">Figure 2
<p>The observational geometry for space-based terahertz radar in detecting space targets.</p>
Full article ">Figure 3
<p>Geometry projection diagram of ISAR imaging.</p>
Full article ">Figure 4
<p>Parabolic antenna imaging characteristics. (<b>a</b>) Three typical observation apertures. (<b>b</b>) Scattering intensity versus azimuth angle. (<b>c</b>) The specular point. (<b>d</b>) The edge pair-points. (<b>e</b>) The ellipse arc.</p>
Full article ">Figure 5
<p>Satellite CAD model with 5 main scattering components (<b>left</b>) and its geometry and size (<b>right</b>).</p>
Full article ">Figure 6
<p>Imaging results and corresponding CAD under three typical observation apertures.</p>
Full article ">Figure 7
<p>Structure diagram of Improved-YOLOv8.</p>
Full article ">Figure 8
<p>Structure diagram of Dyhead.</p>
Full article ">Figure 9
<p>The training samples under different apertures.</p>
Full article ">Figure 10
<p>The distribution of bounding boxes within the dataset.</p>
Full article ">Figure 11
<p>The mAP50 (left) and mAP50-95 (right) of different networks in the training set.</p>
Full article ">Figure 12
<p>A comparison of the detection performance of different algorithms on EM data.</p>
Full article ">Figure 13
<p>mAP50 and mAP50-95 of different networks.</p>
Full article ">Figure 14
<p>PR curves for three different objects.</p>
Full article ">Figure 15
<p>Anechoic chamber experiment and satellite mock-up presentation. (<b>a</b>) Terahertz radar technology system. (<b>b</b>) Satellite model for anechoic chamber experiment.</p>
Full article ">Figure 16
<p>Comparison of performance between different networks on anechoic chamber data.</p>
Full article ">
24 pages, 687 KiB  
Article
MtAD-Net: Multi-Threshold Adaptive Decision Net for Unsupervised Synthetic Aperture Radar Ship Instance Segmentation
by Junfan Xue, Junjun Yin and Jian Yang
Remote Sens. 2025, 17(4), 593; https://doi.org/10.3390/rs17040593 - 9 Feb 2025
Viewed by 501
Abstract
In synthetic aperture radar (SAR) images, pixel-level Ground Truth (GT) is a scarce resource compared to Bounding Box (BBox) annotations. Therefore, exploring the use of unsupervised instance segmentation methods to convert BBox-level annotations into pixel-level GT holds great significance in the SAR field. [...] Read more.
In synthetic aperture radar (SAR) images, pixel-level Ground Truth (GT) is a scarce resource compared to Bounding Box (BBox) annotations. Therefore, exploring the use of unsupervised instance segmentation methods to convert BBox-level annotations into pixel-level GT holds great significance in the SAR field. However, previous unsupervised segmentation methods fail to perform well on SAR images due to the presence of speckle noise, low imaging accuracy, and gradual pixel transitions at the boundaries between targets and background, resulting in unclear edges. In this paper, we propose a Multi-threshold Adaptive Decision Network (MtAD-Net), which is capable of segmenting SAR ship images under unsupervised conditions and demonstrates good performance. Specifically, we design a Multiple CFAR Threshold-extraction Module (MCTM) to obtain a threshold vector by a false alarm rate vector. A Local U-shape Feature Extractor (LUFE) is designed to project each pixel of SAR images into a high-dimensional feature space, and a Global Vision Transformer Encoder (GVTE) is designed to obtain global features, and then, we use the global features to obtain a probability vector, which is the probability of each CFAR threshold. We further propose a PLC-Loss to adaptively reduce the feature distance of pixels of the same category and increase the feature distance of pixels of different categories. Moreover, we designed a label smoothing module to denoise the result of MtAD-Net. Experimental results on the dataset show that our MtAD-Net outperforms traditional and existing deep learning-based unsupervised segmentation methods in terms of pixel accuracy, kappa coefficient, mean intersection over union, frequency weighted intersection over union, and F1-Score. Full article
Show Figures

Figure 1

Figure 1
<p>The overall architecture. In MCTM, thresholds corresponding to different false alarm rates are extracted. The LUFE module maps each pixel in the image to a high-dimensional feature space. The GVTE module employs a Vision Transformer structure to extract global features and maps these global features to the probabilities corresponding to the MCTM output thresholds. We use the designed loss function to update the weights of LUFE and GVTE.The inner product of the output vectors from MTCM and GVTE modules serves as the segmentation threshold for the input image. After segmentation, the result undergoes smoothing through the label smoothing module to obtain the final segmentation result.</p>
Full article ">Figure 2
<p>The image is divided into target region and noise region.</p>
Full article ">Figure 3
<p>Illustration of the Local U-shape Feature Extractor. (<b>a</b>) The overall structure of the LUFE module. After the input image passes through three Down modules and three Up modules, the feature map of the image is obtained. (<b>b</b>) The specific structure of the Down module. (<b>c</b>) The specific structure of the Up module.</p>
Full article ">Figure 4
<p>Illustration of the Global Vision Transformer Encoder. (<b>a</b>) The overall structure of the GVTE module. After Patch and Position Embedding, we feed it into Transformer Encoder, the global features of the image can be obtained. After MLP and SoftMax, global features can be mapped to the probabilities of each CFAR threshold. (<b>b</b>) The specific structure of Transformer Encoder.</p>
Full article ">Figure 5
<p>Schematic diagram of the role of the PLC-Loss function. (<b>a</b>) Distribution of different categories of pixels in the feature space before training. (<b>b</b>) Distribution of different categories of pixels in feature space after training.</p>
Full article ">Figure 6
<p>Qualitative analysis achieved by different unsupervised segmentation methods on small-size ship in the SSDD dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 7
<p>Qualitative analysis achieved by different unsupervised segmentation methods on small-size ship in the HRSID dataset. (<b>a</b>) Input image (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 8
<p>Qualitative analysis achieved by different unsupervised segmentation methods on large ship in the SSDD dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">Figure 9
<p>Qualitative analysis achieved by different unsupervised segmentation methods on large ship in the HRSID dataset. (<b>a</b>) Input image, (<b>b</b>) GT, (<b>c</b>) CFAR, (<b>d</b>) OTSU, (<b>e</b>) Kim-Net, (<b>f</b>) PiCIE, (<b>g</b>) CDA-SAR, (<b>h</b>) IDUDL, (<b>i</b>) Ours. The false negative areas and false positive areas are highlighted by green and red.</p>
Full article ">
21 pages, 25589 KiB  
Article
Robust and Efficient SAR Ship Detection: An Integrated Despecking and Detection Framework
by Yulin Chen, Yanyun Shen, Chi Duan, Zhipan Wang, Zewen Mo, Yingyu Liang and Qingling Zhang
Remote Sens. 2025, 17(4), 580; https://doi.org/10.3390/rs17040580 - 8 Feb 2025
Viewed by 275
Abstract
Deep-learning-based ship detection methods in Synthetic Aperture Radar (SAR) imagery are a current research hotspot. However, these methods rely on high-quality images as input, and in practical applications, SAR images are interfered with by speckle noise, leading to a decrease in image quality [...] Read more.
Deep-learning-based ship detection methods in Synthetic Aperture Radar (SAR) imagery are a current research hotspot. However, these methods rely on high-quality images as input, and in practical applications, SAR images are interfered with by speckle noise, leading to a decrease in image quality and thus affecting detection accuracy. To address this problem, we propose a unified framework for ship detection that incorporates a despeckling module into the object detection network. This integration is designed to enhance the detection performance, even with low-quality SAR images that are affected by speckle noise. Secondly, we propose a Multi-Scale Window Swin Transformer module. This module is adept at improving image quality by effectively capturing both global and local features of the SAR images. Additionally, recognizing the challenges associated with the scarcity of labeled data in practical scenarios, we employ an unlabeled distillation learning method to train our despeckling module. This technique avoids the need for extensive manual labeling and making efficient use of unlabeled data. We have tested the robustness of our method using public SAR datasets, including SSDD and HRSID, as well as a newly constructed dataset, the RSSDD. The results demonstrate that our method not only achieves a state-of-the-art performance but also excels in conditions with low signal-to-noise ratios. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Overall framework of our method.</p>
Full article ">Figure 2
<p>Framework of teacher–student model, where * denotes the multiplication sign.</p>
Full article ">Figure 3
<p>Structure of Multi-scale Window Swin Transformer (MSW Swin Transformer) Module.</p>
Full article ">Figure 4
<p>Mobile window reorganization.</p>
Full article ">Figure 5
<p>Structure of LFE module, where * denotes the multiplication sign.</p>
Full article ">Figure 6
<p>Ship detection model with noise robustness based on YOLOv5.</p>
Full article ">Figure 7
<p>Samples from RSSDD.</p>
Full article ">Figure 8
<p>The detection results of our method based on YOLOv5 with different noise parameters L = [1,2,3,4]. (<b>a</b>) Ground truth, (<b>b</b>) L = 4, (<b>c</b>) L = 3, (<b>d</b>) L = 2, (<b>e</b>) L = 1. The green box represents ground truth box and the red box represents prediction box.</p>
Full article ">Figure 9
<p>Detection results with Centernet. (<b>a</b>) Ground truth, detection results of (<b>b</b>) baseline model, (<b>c</b>) SARCNN, (<b>d</b>) SARDRN, (<b>e</b>) Gamma filter, (<b>f</b>) Lee filter, (<b>g</b>) Frost filter, and (<b>h</b>) proposed method. The green box represents ground truth box and the red box represents prediction box.</p>
Full article ">Figure 10
<p>Visualization of object detection algorithm in SSDD and RSSDD. (<b>a</b>) Ground truth, (<b>b</b>) baseline, (<b>c</b>) Gamma, (<b>d</b>) ours. The green box represents ground truth box and the red box represents prediction box.</p>
Full article ">
13 pages, 2045 KiB  
Article
A Hardware Accelerator for Real-Time Processing Platforms Used in Synthetic Aperture Radar Target Detection Tasks
by Yue Zhang, Yunshan Tang, Yue Cao and Zhongjun Yu
Micromachines 2025, 16(2), 193; https://doi.org/10.3390/mi16020193 - 7 Feb 2025
Viewed by 425
Abstract
The deep learning object detection algorithm has been widely applied in the field of synthetic aperture radar (SAR). By utilizing deep convolutional neural networks (CNNs) and other techniques, these algorithms can effectively identify and locate targets in SAR images, thereby improving the accuracy [...] Read more.
The deep learning object detection algorithm has been widely applied in the field of synthetic aperture radar (SAR). By utilizing deep convolutional neural networks (CNNs) and other techniques, these algorithms can effectively identify and locate targets in SAR images, thereby improving the accuracy and efficiency of detection. In recent years, achieving real-time monitoring of regions has become a pressing need, leading to the direct completion of real-time SAR image target detection on airborne or satellite-borne real-time processing platforms. However, current GPU-based real-time processing platforms struggle to meet the power consumption requirements of airborne or satellite applications. To address this issue, a low-power, low-latency deep learning SAR object detection algorithm accelerator was designed in this study to enable real-time target detection on airborne and satellite SAR platforms. This accelerator proposes a Process Engine (PE) suitable for multidimensional convolution parallel computing, making full use of Field-Programmable Gate Array (FPGA) computing resources to reduce convolution computing time. Furthermore, a unique memory arrangement design based on this PE aims to enhance memory read/write efficiency while applying dataflow patterns suitable for FPGA computing to the accelerator to reduce computation latency. Our experimental results demonstrate that deploying the SAR object detection algorithm based on Yolov5s on this accelerator design, mounted on a Virtex 7 690t chip, consumes only 7 watts of dynamic power, achieving the capability to detect 52.19 512 × 512-sized SAR images per second. Full article
(This article belongs to the Section E:Engineering and Technology)
Show Figures

Figure 1

Figure 1
<p>The structure of the entire accelerator.</p>
Full article ">Figure 2
<p>The specific breakdown of the convolution operation.</p>
Full article ">Figure 3
<p>The data input order of the PE. (<b>a</b>) The data input corresponding to Loop6. (<b>b</b>) The data input corresponding to Loop5. (<b>c</b>) The data input corresponding to Loop4. (<b>d</b>) The data input corresponding to Loop3. (<b>e</b>) The data input corresponding to Loop2. (<b>f</b>) The data input corresponding to Loop1.</p>
Full article ">Figure 4
<p>The hardware structure of the PE.</p>
Full article ">Figure 5
<p>The struct of Buffer. (<b>a</b>) The struct of FmBuffer. (<b>b</b>) The struct of WtBuffer.</p>
Full article ">Figure 6
<p>The memory arrangements. (<b>a</b>) The memory arrangement of FeatureMap. (<b>b</b>) The memory arrangement of Weight.</p>
Full article ">Figure 7
<p>The streaming computing mode.</p>
Full article ">Figure 8
<p>The figure of dynamic power.</p>
Full article ">
28 pages, 11323 KiB  
Article
Polarimetric SAR Ship Detection Using Context Aggregation Network Enhanced by Local and Edge Component Characteristics
by Canbin Hu, Hongyun Chen, Xiaokun Sun and Fei Ma
Remote Sens. 2025, 17(4), 568; https://doi.org/10.3390/rs17040568 - 7 Feb 2025
Viewed by 323
Abstract
Polarimetric decomposition methods are widely used in polarimetric Synthetic Aperture Radar (SAR) data processing for extracting scattering characteristics of targets. However, polarization SAR methods for ship detection still face challenges. The traditional constant false alarm rate (CFAR) detectors face sea clutter modeling and [...] Read more.
Polarimetric decomposition methods are widely used in polarimetric Synthetic Aperture Radar (SAR) data processing for extracting scattering characteristics of targets. However, polarization SAR methods for ship detection still face challenges. The traditional constant false alarm rate (CFAR) detectors face sea clutter modeling and parameter estimation problems in ship detection, which is difficult to adapt to the complex background. In addition, neural network-based detection methods mostly rely on single polarimetric-channel scattering information and fail to fully explore the polarization properties and physical scattering laws of ships. To address these issues, this study constructed two novel characteristics: a helix-scattering enhanced (HSE) local component and a multi-scattering intensity difference (MSID) edge component, which are specifically designed to describe ship scattering characteristics. Based on the characteristic differences of different scattering components in ships, this paper designs a context aggregation network enhanced by local and edge component characteristics to fully utilize the scattering information of polarized SAR data. With the powerful feature extraction capability of a convolutional neural network, the proposed method can significantly enhance the distinction between ships and the sea. Further analysis shows that HSE is able to capture structural information about the target, MSID can increase ship–sea separation capability, and an HV channel retains more detailed information. Compared with other decomposition models, the proposed characteristic combination model performs well in complex backgrounds and can distinguish ship from sea more effectively. The experimental results show that the proposed method achieves a detection precision of 93.6% and a recall rate of 91.5% on a fully polarized SAR dataset, which are better than other popular network algorithms, verifying the reasonableness and superiority of the method. Full article
Show Figures

Figure 1

Figure 1
<p>Ship scattering characteristics in four component decomposition.</p>
Full article ">Figure 2
<p>Enhancement comparison before (<b>a</b>) and after (<b>b</b>) the difference.</p>
Full article ">Figure 3
<p>Structural diagram of context aggregation network based on local and edge component feature enhancement.</p>
Full article ">Figure 4
<p>Scattering Structure Feature Extraction Network.</p>
Full article ">Figure 5
<p>Detailed view of DCNblock module.</p>
Full article ">Figure 6
<p>Structure of the CAM.</p>
Full article ">Figure 7
<p>Low-Level Feature Guided Balanced Fusion Network for PolSAR.</p>
Full article ">Figure 8
<p>Comparison of extracted characteristics from RADARSAT-2 data. (<b>a1</b>,<b>a2</b>) Pauli pseudocolor maps; (<b>b1</b>,<b>b2</b>) HSE; (<b>c1</b>,<b>c2</b>) MSID; (<b>d1</b>,<b>d2</b>) HH; and (<b>e1</b>,<b>e2</b>) HV.</p>
Full article ">Figure 9
<p>Comparison of extracted characteristics from AIRSAR data. (<b>a1</b>,<b>a2</b>) Pauli pseudocolor maps; (<b>b1</b>,<b>b2</b>) HSE; (<b>c1</b>,<b>c2</b>) MSID; (<b>d1</b>,<b>d2</b>) HH; and (<b>e1</b>,<b>e2</b>) HV.</p>
Full article ">Figure 10
<p>Comparison of extracted characteristics from UAVSAR data. (<b>a1</b>,<b>a2</b>) Pauli pseudocolor maps; (<b>b1</b>,<b>b2</b>) HSE; (<b>c1</b>,<b>c2</b>) MSID; (<b>d1</b>,<b>d2</b>) HH; and (<b>e1</b>,<b>e2</b>) HV.</p>
Full article ">Figure 11
<p>3D scatter plots of ship and sea characteristics. (<b>a</b>) Pauli pseudocolor map; (<b>b</b>) Pauli decomposition 3D scatter plot; (<b>c</b>) Freeman–Durden decomposition 3D scatter plot; and (<b>d</b>) proposed characteristics 3D scatter plot.</p>
Full article ">Figure 12
<p>Distribution of target pixel sizes.</p>
Full article ">Figure 13
<p>Comparison of ship detection results under different polarimetric characteristic combinations. Green rectangles indicate the ground truth, red rectangles indicate the detected results, blue circles indicate the false alarms, and orange circles indicate the missed detections. (<b>a</b>) Ground truth; (<b>b</b>) Pauli components; (<b>c</b>) Freeman–Durden components; (<b>d</b>) Proposed method.</p>
Full article ">Figure 14
<p>Comparison of feature maps under different backbone networks. (<b>a</b>) Pauli image; (<b>b</b>) feature map generated by the backbone network constructed with traditional convolutional blocks; (<b>c</b>) feature map generated by the proposed backbone network employing deformable convolutional blocks.</p>
Full article ">Figure 15
<p>Comparison of ship detection results under different network modules. Green rectangles indicate the ground truth, red rectangles indicate the detected results, blue circles indicate the false alarms, and orange circles indicate the missed detections. (<b>a</b>) Ground truth; (<b>b</b>) CAM only; (<b>c</b>) DCNblock only; (<b>d</b>) both DCNblock and CAM.</p>
Full article ">Figure 16
<p>Comparison of vessel detection results under different networks. Green rectangles indicate the ground truth, red rectangles indicate the detected results, blue circles indicate the false alarms, and orange circles indicate the missed detections. (<b>a</b>) Ground truth, (<b>b</b>) RetinaNet, (<b>c</b>) CenterNet, (<b>d</b>) Faster-RCNN, (<b>e</b>) YOLOv5, (<b>f</b>) YOLOv8, (<b>g</b>) MobileNet, (<b>h</b>) Proposed method.</p>
Full article ">
21 pages, 53374 KiB  
Article
FloodKAN: Integrating Kolmogorov–Arnold Networks for Efficient Flood Extent Extraction
by Cong Wang, Xiaohan Zhang and Liwei Liu
Remote Sens. 2025, 17(4), 564; https://doi.org/10.3390/rs17040564 - 7 Feb 2025
Viewed by 480
Abstract
Flood events are among the most destructive natural catastrophes worldwide and pose serious threats to socioeconomic systems, ecological environments, and the safety of human life and property. With the advancement of remote sensing technology, synthetic aperture radar (SAR) has provided new means for [...] Read more.
Flood events are among the most destructive natural catastrophes worldwide and pose serious threats to socioeconomic systems, ecological environments, and the safety of human life and property. With the advancement of remote sensing technology, synthetic aperture radar (SAR) has provided new means for flood monitoring. However, traditional methods have limitations when dealing with high noise levels and complex terrain backgrounds. To address this issue, in this study, we adopt an improved U-Net model incorporating the Kolmogorov–Arnold Network (KAN), referred to as UKAN, for the efficient extraction of flood inundation extents from multisource remote sensing data. UKAN integrates the efficient nonlinear mapping capabilities of KAN layers with the multiscale feature fusion mechanism of U-Net, enabling better capturing of complex nonlinear relationships and global features. Experiments were conducted on the C2S-MS Floods and MMFlood datasets, and the results indicate that the UKAN model outperforms traditional models in terms of metrics such as the intersection over union (IoU), precision, recall, and F1 score. On the C2S-MS Floods dataset and the MMFlood dataset, UKAN achieves IoUs of 87.95% and 78.31%, respectively, representing improvements of approximately 3.5 and three percentage points, respectively, over those of the traditional U-Net. Moreover, the model has significant advantages in terms of parameter efficiency and computational efficiency. These findings suggest that the UKAN model possesses greater accuracy and robustness in flood inundation area extraction tasks, which is highly important for increasing the monitoring and early warning capabilities of flood disasters. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Example of C2S-MS Floods data.</p>
Full article ">Figure 2
<p>Example of MMFlood data.</p>
Full article ">Figure 3
<p>Overall architecture of flood UKAN.</p>
Full article ">Figure 4
<p>Structures of the convolution block and KAN block.</p>
Full article ">Figure 5
<p>Structure of the KAN layer.</p>
Full article ">Figure 6
<p>Comparison of results for C2S-MS Floods.</p>
Full article ">Figure 7
<p>Comparison of results for MMFlood.</p>
Full article ">Figure 8
<p>Comparative analysis of failure cases in flood inundation prediction across models.</p>
Full article ">Figure 9
<p>Comparison of classification performance for C2S-MS Floods.</p>
Full article ">Figure 10
<p>Comparison of classification performance for MMFlood.</p>
Full article ">Figure 11
<p>Evaluation of segmentation models: metrics, parameters, and computational complexity (input size = (16, 4, 256, 256)).</p>
Full article ">
71 pages, 4216 KiB  
Review
Advances in Remote Sensing and Deep Learning in Coastal Boundary Extraction for Erosion Monitoring
by Marc-André Blais and Moulay A. Akhloufi
Geomatics 2025, 5(1), 9; https://doi.org/10.3390/geomatics5010009 - 6 Feb 2025
Viewed by 435
Abstract
Erosion is a critical geological process that degrades soil and poses significant risks to human settlements and natural habitats. As climate change intensifies, effective coastal erosion management and prevention have become essential for our society and the health of our planet. Given the [...] Read more.
Erosion is a critical geological process that degrades soil and poses significant risks to human settlements and natural habitats. As climate change intensifies, effective coastal erosion management and prevention have become essential for our society and the health of our planet. Given the vast extent of coastal areas, erosion management efforts must prioritize the most vulnerable and critical regions. Identifying and prioritizing these areas is a complex task that requires the accurate monitoring and forecasting of erosion and its potential impacts. Various tools and techniques have been proposed to assess the risks, impacts and rates of coastal erosion. Specialized methods, such as the Coastal Vulnerability Index, have been specifically designed to evaluate the susceptibility of coastal areas to erosion. Coastal boundaries, a critical factor in coastal erosion monitoring, are typically extracted from remote sensing images. Due to the extensive scale of coastal areas and the complexity of the data, manually extracting coastal boundaries is challenging. Recently, artificial intelligence, particularly deep learning, has emerged as a promising and essential tool for this task. This review provides an in-depth analysis of remote sensing and deep learning for extracting coastal boundaries to assist in erosion monitoring. Various remote sensing imaging modalities (optical, thermal, radar), platforms (satellites, drones) and datasets are first presented to provide the context for this field. Artificial intelligence and its associated metrics are then discussed, followed by an exploration of deep learning algorithms for extracting coastal boundaries. The presented algorithms range from basic convolutional networks to encoder–decoder architectures and attention mechanisms. An overview of how these extracted boundaries and other deep learning algorithms can be utilized for monitoring coastal erosion is also provided. Finally, the current gaps, limitations and potential future directions in this field are identified. This review aims to offer critical insights into the future of erosion monitoring and management through deep learning-based boundary extraction. Full article
Show Figures

Figure 1

Figure 1
<p>Overview of the review structure.</p>
Full article ">Figure 2
<p>Electromagnetic spectrum.</p>
Full article ">Figure 3
<p>Results of pan-sharpening.</p>
Full article ">Figure 4
<p>Visualization of composites: FC1 (NIR, Red, Green), FC2 (NIR, SWIR2, Red) and true color (Red, Green, Blue).</p>
Full article ">Figure 5
<p>Examples of indices for the same region.</p>
Full article ">Figure 6
<p>Examples of datasets: RGB in the first row, the corresponding masks in the second.</p>
Full article ">Figure 7
<p>(<b>a</b>) AI subsets showing the hierarchy of AI, ML and DL. (<b>b</b>) A simple neural network diagram explaining its components.</p>
Full article ">Figure 8
<p>Differences in algorithms.</p>
Full article ">Figure 9
<p>UNet structure.</p>
Full article ">Figure 10
<p>DeepLabV3+ structure.</p>
Full article ">Figure 11
<p>DANet-SMIW structure.</p>
Full article ">Figure 12
<p>SRMA structure.</p>
Full article ">Figure 13
<p>LaeNet structure.</p>
Full article ">Figure 14
<p>CSAFNet structure.</p>
Full article ">
Back to TopTop