Nothing Special   »   [go: up one dir, main page]

You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (4,178)

Search Parameters:
Keywords = synthetic aperture radar (SAR)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
12 pages, 4167 KiB  
Article
Development of a Deployable Reflector Antenna for the Synthetic Aperture Radar Satellite, Part 2: Manufacturing and Qualification of the Main Reflector Using a Honeycomb Sandwich Composite Structure
by Dong-Geon Kim, Hyun-Guk Kim, Dong-Yeon Kim, Ryoon-Ho Do, Kyung-Rae Koo and Young-Joon Yu
Appl. Sci. 2024, 14(23), 11273; https://doi.org/10.3390/app142311273 - 3 Dec 2024
Abstract
A deployable reflector antenna (DR-A) is a structure that can be stored in a large-diameter Synthetic Aperture Radar (SAR) antenna and be mounted onto a launch vehicle. Considering the performance of the launch vehicle, it is necessary to develop a lightweight, high-performance antenna [...] Read more.
A deployable reflector antenna (DR-A) is a structure that can be stored in a large-diameter Synthetic Aperture Radar (SAR) antenna and be mounted onto a launch vehicle. Considering the performance of the launch vehicle, it is necessary to develop a lightweight, high-performance antenna structure. The solid-type deployable reflector antenna is composed of a number of unit main reflectors. To reduce the weight of the antenna, a lightweight main reflector must be developed. In this paper, following “Development of Deployable Reflector Antenna for the SAR Satellite (Part 1)”, the manufacturing and qualification of the main reflector using honeycomb sandwich composites are described. Four types of composite main reflectors were manufactured with variables in the manufacturing process. The manufacturing variables include the curing process of the structure, the application of an adhesive film between the sheet and the core, and the venting path inside of the sandwich core. After manufacturing the main reflector, we performed weight measurements, non-destructive testing (NDT), surface error measurement using a Coordinate Measurement Machine (CMM), and modal testing for each type of composite main reflector. Through the research and development process, we found that a perforated hole is necessary when excluding the adhesive film during bonding of an aramid core and a CFRP sheet, and a lightweight composite reflector could be developed through this process. We selected the main reflector with the best performance and developed a composite main reflector that can be applied to satellites. Full article
(This article belongs to the Section Aerospace Science and Engineering)
Show Figures

Figure 1

Figure 1
<p>Design of composite main reflector under density of honeycomb core.</p>
Full article ">Figure 2
<p>Modal analysis results for deployable reflector antenna (low-density core).</p>
Full article ">Figure 3
<p>Stacking of composite main reflector.</p>
Full article ">Figure 4
<p>Pressure distribution during curing without venting path.</p>
Full article ">Figure 5
<p>De-bonding between face sheet (CFRP) and honeycomb core when curing without venting path.</p>
Full article ">Figure 6
<p>Pressure distribution during curing with venting path.</p>
Full article ">Figure 7
<p>Perforated holes in aramid honeycomb core.</p>
Full article ">Figure 8
<p>Additional laminate patch-up at the interface section of the main reflector.</p>
Full article ">Figure 9
<p>Manufacturing of sandwich composite main reflector.</p>
Full article ">Figure 10
<p>Comparison of the surface of the main reflector based on curing conditions.</p>
Full article ">Figure 11
<p>Results of NDT for main reflectors.</p>
Full article ">Figure 12
<p>Coordinate measurement configuration considering gravity.</p>
Full article ">Figure 13
<p>Test configuration and result of modal test on reflector.</p>
Full article ">
20 pages, 12422 KiB  
Article
LHSDNet: A Lightweight and High-Accuracy SAR Ship Object Detection Algorithm
by Dahai Dai, Hao Wu, Yue Wang and Penghui Ji
Remote Sens. 2024, 16(23), 4527; https://doi.org/10.3390/rs16234527 - 3 Dec 2024
Viewed by 105
Abstract
At present, the majority of deep learning-based ship object detection algorithms concentrate predominantly on enhancing recognition accuracy, often overlooking the complexity of the algorithm. These complex algorithms demand significant computational resources, making them unsuitable for deployment on resource-constrained edge devices, such as airborne [...] Read more.
At present, the majority of deep learning-based ship object detection algorithms concentrate predominantly on enhancing recognition accuracy, often overlooking the complexity of the algorithm. These complex algorithms demand significant computational resources, making them unsuitable for deployment on resource-constrained edge devices, such as airborne and spaceborne platforms, thereby limiting their practicality. With the purpose of alleviating this problem, a lightweight and high-accuracy synthetic aperture radar (SAR) ship image detection network (LHSDNet) is proposed. Initially, GhostHGNetV2 was utilized as the feature extraction network, and the calculation amount of the network was reduced by GhostConv. Next, a lightweight feature fusion network was designed to combine shallow and deep features through lightweight convolutions, effectively preserving more information while minimizing computational requirements. Lastly, the feature extraction module was integrated through parameter sharing, and the detection head was lightweight to save computing resources further. The results from our experiments demonstrate that the proposed LHSDNet model increases mAP50 by 0.7% in comparison to the baseline model. Additionally, it illustrates a pronounced decrease in parameter count, computational demand, and model file size by 48.33%, 51.85%, and 41.26%, respectively, when contrasted with the baseline model. LHSDNet achieves a balance between precision and computing resources, rendering it more appropriate for edge device implementation. Full article
Show Figures

Figure 1

Figure 1
<p>The overall architecture of YOLOv8.</p>
Full article ">Figure 2
<p>Overall structure of LHSDNet.</p>
Full article ">Figure 3
<p>The structure of HGStem and HGBlock.</p>
Full article ">Figure 4
<p>The structure of GhostConv.</p>
Full article ">Figure 5
<p>The structure of DWConv.</p>
Full article ">Figure 6
<p>Lightweight feature fusion network architecture.</p>
Full article ">Figure 7
<p>Reparameterization method.</p>
Full article ">Figure 8
<p>Comparison of detection networks: (<b>a</b>) is the detection network of YOLOv8n and (<b>b</b>) is the detection network of LHSDNet with shared parameters.</p>
Full article ">Figure 9
<p>The architecture of PConv.</p>
Full article ">Figure 10
<p>Part of the images in the dataset.</p>
Full article ">Figure 11
<p>Detection results of LHSDNet.</p>
Full article ">Figure 12
<p>Comparison of heat maps between LHSDNet and YOLOv8n.</p>
Full article ">Figure 13
<p>Visualization of detection performance of different algorithms in inshore scenarios.</p>
Full article ">Figure 14
<p>Visualization of detection performance of different algorithms in offshore scenarios.</p>
Full article ">Figure 15
<p>Performance comparison between LHSDNet and YOLOv8n during training.</p>
Full article ">
21 pages, 9860 KiB  
Article
Uncertainty Analysis of Forest Aboveground Carbon Stock Estimation Combining Sentinel-1 and Sentinel-2 Images
by Bo Qiu, Sha Li, Jun Cao, Jialong Zhang, Kun Yang, Kai Luo, Kai Huang and Xinzhou Jiang
Forests 2024, 15(12), 2134; https://doi.org/10.3390/f15122134 - 2 Dec 2024
Viewed by 311
Abstract
Accurate estimation of forest aboveground carbon stock (AGC) is essential for understanding carbon accounting and climate change. In previous studies, the extracted factors, such as spectral textures, vegetation indices, and textural features, were used to estimate the AGC. However, few studies examined how [...] Read more.
Accurate estimation of forest aboveground carbon stock (AGC) is essential for understanding carbon accounting and climate change. In previous studies, the extracted factors, such as spectral textures, vegetation indices, and textural features, were used to estimate the AGC. However, few studies examined how different factors affect estimation accuracy in detail. Meanwhile, there are also many uncertainties in the collection and processing of the field data. To quantify the various uncertainties in the process of AGC estimation, we used the random forest (RF) to establish estimation models based on field data and Sentinel-1/2 images in Shangri-La. The models included the band information model (BIM), the vegetation index model (VIM), the texture information model (TIM), the Sentinel-2 factor model (S-2M), and the Sentinel-1/2 factor model (S-1/2M). Then, uncertainties resulting from the plot scale and estimation models were calculated using error equations. Our goal is to analyze the influence of different factors on AGC estimation and to assess the uncertainty of plot scale and estimation models quantitatively. The results showed that (1) the uncertainty of the measurement was 3.02%, while the error of the monocarbon stock model was the main uncertainty at the plot scale, which was 9.09%; (2) the BIM had the lowest accuracy (R2 = 0.551) and the highest total uncertainty (22.29%); by gradually introducing different factors in the process of modeling, the accuracies improved significantly (VIM: R2 = 0.688, TIM: R2 = 0.715, S-2M: R2 = 0.826), and the total uncertainty decreased to some extent (VIM: 14.12%, TIM: 12.56%, S-2M: 10.79%); (3) the S-1/2M with the introduction of Sentinel-1 synthetic aperture radar (SAR) data has the highest accuracy (R2 = 0.872) and the lowest total uncertainty (8.43%). The inaccuracy of spectral features is highest, followed by vegetation indices, while textural features have the lowest inaccuracy. Uncertainty in the remote-sensing-based estimation model remains a significant source of uncertainty compared to the plot scale. Even though the uncertainty at the plot scale is relatively small, this error should not be ignored. The uncertainty in the estimation process could be further reduced by improving the precision of the measurement and the fitting of the monocarbon stock estimation model. Full article
(This article belongs to the Special Issue Modeling Aboveground Forest Biomass: New Developments)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Location of Yunnan Province in China; (<b>b</b>) location of Shangri-La in Yunnan Province; (<b>c</b>) distribution of <span class="html-italic">Pinus densata</span> and field plots; and (<b>d</b>) elevation.</p>
Full article ">Figure 2
<p>Overview of the methodological approach.</p>
Full article ">Figure 3
<p>(<b>a</b>) Scatterplot based on the band information for modeling; (<b>b</b>) scatterplot based on the vegetation index for modeling; (<b>c</b>) scatterplot based on the texture information for modeling; (<b>d</b>) scatterplot based on the Sentinel-2 factor for modeling; and (<b>e</b>) scatterplot based on the Sentinel-1/2 factor for modeling.</p>
Full article ">Figure 4
<p>(<b>a</b>) The distribution of <span class="html-italic">R</span><sup>2</sup> values for different models and (<b>b</b>) the distribution of <span class="html-italic">RMSE</span> values for different models.</p>
Full article ">Figure 5
<p>(<b>a</b>) Relationship between <span class="html-italic">DBH</span> and standard deviation of measurement error and (<b>b</b>) relationship between <span class="html-italic">H</span> and measurement error.</p>
Full article ">Figure 6
<p>(<b>a</b>) Three-dimensional scatterplot of the binary carbon stock model for <span class="html-italic">Pinus densata</span> and (<b>b</b>) line plot of errors and accuracy for the model.</p>
Full article ">Figure 7
<p>(<b>a</b>) Error equations for the BIM, VIM, and TIM and (<b>b</b>) error equations for the S-2M and S-1/2M.</p>
Full article ">Figure 8
<p>(<b>a</b>) Difference in each field plot for BIM, VIM, and TIM and (<b>b</b>) difference in each field plot for the S-2M and S-1/2M.</p>
Full article ">Figure 9
<p>Uncertainty of different remote-sensing-based models.</p>
Full article ">
13 pages, 4906 KiB  
Technical Note
An Extended Omega-K Algorithm for Automotive SAR with Curved Path
by Ping Guo, Chao Li, Haolan Li, Yuchen Luan, Anyi Wang, Rongshu Wang and Shiyang Tang
Remote Sens. 2024, 16(23), 4508; https://doi.org/10.3390/rs16234508 - 1 Dec 2024
Viewed by 254
Abstract
Automotive millimeter-wave (MMW) synthetic aperture radar (SAR) systems can achieve high-resolution images of detection areas, providing environmental perceptions that facilitate intelligent driving. However, curved path is inevitable in complex urban road environments. Non-uniform spatial sampling, brought about by curved path, leads to cross-coupling [...] Read more.
Automotive millimeter-wave (MMW) synthetic aperture radar (SAR) systems can achieve high-resolution images of detection areas, providing environmental perceptions that facilitate intelligent driving. However, curved path is inevitable in complex urban road environments. Non-uniform spatial sampling, brought about by curved path, leads to cross-coupling and spatial variation deteriorates greatly, significantly impacting the imaging results. To deal with these issues, we developed an Extended Omega-K Algorithm (EOKA) for an automotive SAR with a curved path. First, an equivalent range model was constructed based on the relationship between the range history and Doppler frequency. Then, using azimuth time mapping, the echo data was reconstructed with a form similar to that of a uniform linear case. As a result, an analytical two-dimensional (2D) spectrum was easily derived without using of the method of series reversion (MSR) that could be exploited for EOKA. The results from the parking lot, open road, and obstacle experimental scenes demonstrate the performance and feasibility of an MMW SAR for environmental perception. Full article
Show Figures

Figure 1

Figure 1
<p>Geometry of automotive SAR with curved path.</p>
Full article ">Figure 2
<p>Real data of INS and fitting results. (<b>a</b>) X, (<b>b</b>) Y and (<b>c</b>) Z.</p>
Full article ">Figure 3
<p>The phase errors. (<b>a</b>) Fitting, (<b>b</b>) Equation (2).</p>
Full article ">Figure 4
<p>Range history reconstruction diagram.</p>
Full article ">Figure 5
<p>Flowchart of the imaging algorithm.</p>
Full article ">Figure 6
<p>The image scenes. (<b>a</b>) simulated scene; (<b>b</b>) experimental scene.</p>
Full article ">Figure 7
<p>The IRF of three targets. (<b>a</b>) OKA, (<b>b</b>) EOKA, (<b>c</b>) FFBPA.</p>
Full article ">Figure 8
<p>The optical image of obstacle scene.</p>
Full article ">Figure 9
<p>Obstacle focused images. (<b>a</b>) EOKA, (<b>b</b>) OKA.</p>
Full article ">Figure 10
<p>Focused image. (<b>a</b>) Parking lot scene; (<b>b</b>) Open road scene.</p>
Full article ">
22 pages, 4202 KiB  
Article
Spatial-Temporal Variations and Severity of the 2020 Catastrophic Floods in the Yangtze River Basin from Sentinel-1 SAR Data
by Minmin Huang, Ying Wang, Shuanggen Jin and Qingxiang Chen
Water 2024, 16(23), 3445; https://doi.org/10.3390/w16233445 - 29 Nov 2024
Viewed by 290
Abstract
Flood is one of the most frequent natural disasters in the Yangtze River Basin. Flood risk evaluation is of great social significance, especially for large hydrological systems. Rainfall is both temporal and spatial, influencing surface hydrological activities. The water body range is the [...] Read more.
Flood is one of the most frequent natural disasters in the Yangtze River Basin. Flood risk evaluation is of great social significance, especially for large hydrological systems. Rainfall is both temporal and spatial, influencing surface hydrological activities. The water body range is the final outcome of a flood and can be observed from synthetic aperture radar (SAR) images under any weather condition. A flood severity evaluation model is proposed to quantitatively evaluate the flood based on water body range from area disparity and flood duration. Large hydrological objects usually span a wide range and have significant differences. This results in different initial water areas in each region. This approach addresses the issue through normalization processing. In this paper, Sentinel-1 data are used to extract the temporal water body using the adaptive bimodal method, and the water level data were also incorporated to improve the observation frequency for water area. The flood severity evaluation approach can be used to assess flood risk between any region of large hydrological systems or any flood event, regardless of their regional spatial differences and rainfall duration differences. The results show that: (1) In general, the average water body area in 2020 was 20.40% larger than it was in 2019, and the daily water body areas in 2020 were all greater than the average of 2019 with 71.36% of the days in 2020 having an area greater than the maximum in 2019. The flood severity in 2020 was 1.75 times as much as that of 2019; (2) Reach performance indexes in 2020 were in order of Yueyang (2.21) > Jiujiang (2.04) > Hankou (1.44) > Chizhou (1.32), which were inconsistent with the spatial site; (3) Flood event impact indexes in 2020 were in order as No.2 (1.64) > No.3 (1.61) > No.1 (1.44) > No.4 (1.17) > No.5 (1.15); (4) The flood was more likely the result of cumulative rainfall for 30 days. Full article
Show Figures

Figure 1

Figure 1
<p>Map of the study area: (<b>a</b>) the middle and lower streams of YZRB; (<b>b</b>) research scope of Yueyang reach; (<b>c</b>) research scope of Hankou reach; (<b>d</b>) research scope of Jiujiang reach; (<b>e</b>) research scope of Chizhou reach.</p>
Full article ">Figure 2
<p>Flow chart of water extraction and flood evaluation.</p>
Full article ">Figure 3
<p>(<b>a</b>) Sentinel-1 GRD image of VH polarization in Wuhan on 6 August 2020; (<b>b</b>) Histogram of the image indicating the threshold to retrieve water body.</p>
Full article ">Figure 4
<p>The line charts of rainfall during each flood event: (<b>a</b>) The average daily rainfall; (<b>b</b>) The maximum daily rainfall; (<b>c</b>) The cumulative rainfall during each flood event; (<b>d</b>) The cumulative rainfall for 10 days before the flood event; (<b>e</b>) The cumulative rainfall for 15 days before the flood event; (<b>f</b>) The cumulative rainfall for 30 days before the flood event.</p>
Full article ">Figure 5
<p>Planet images in different reaches: (<b>a</b>) Yueyang reach on 5 August 2020; (<b>b</b>) Yueyang reach on 23 August 2020; (<b>c</b>) Hankou reach on 1 August 2020; (<b>d</b>) Jiujiang reach on 19 August 2020.</p>
Full article ">Figure 6
<p>(<b>a</b>) Increase in average area of each flood event in 2020 compared with the average area during July and August in 2019; (<b>b</b>) Increase in max area of each flood event in 2020 compared with the max area during July and August in 2019.</p>
Full article ">Figure 7
<p>The flood map of four reaches: (<b>a</b>) Yueyang reach; (<b>b</b>) Hankou reach; (<b>c</b>) Jiujiang reach; (<b>d</b>) Chizhou reach; (<b>e</b>–<b>h</b>): Local enlarged drawings of Chizhou reach.</p>
Full article ">Figure 7 Cont.
<p>The flood map of four reaches: (<b>a</b>) Yueyang reach; (<b>b</b>) Hankou reach; (<b>c</b>) Jiujiang reach; (<b>d</b>) Chizhou reach; (<b>e</b>–<b>h</b>): Local enlarged drawings of Chizhou reach.</p>
Full article ">Figure 8
<p>(<b>a</b>) The cumulative rainfall during 10 days before the flood event; (<b>b</b>) The cumulative rainfall for 15 days before the flood event; (<b>c</b>) The cumulative rainfall for 30 days before the flood event; (<b>d</b>) Water body area increases in each flood event in 2020 compared with that in 2019; (<b>e</b>) Flood severity index of each flood event in 2020.</p>
Full article ">
20 pages, 6279 KiB  
Article
Self-Activated Implicit Neural Representation for Synthetic Aperture Radar Images
by Dongshen Han and Chaoning Zhang
Remote Sens. 2024, 16(23), 4473; https://doi.org/10.3390/rs16234473 - 28 Nov 2024
Viewed by 251
Abstract
Image Implicit Neural Representations (INRs) adopt a neural network to learn a continuous function for mapping the pixel coordinates to their corresponding values. This task has gained significant attention for representing images in a continuous manner. Despite substantial progress regarding natural images, there [...] Read more.
Image Implicit Neural Representations (INRs) adopt a neural network to learn a continuous function for mapping the pixel coordinates to their corresponding values. This task has gained significant attention for representing images in a continuous manner. Despite substantial progress regarding natural images, there is little investigation of INRs for Synthetic Aperture Radar (SAR) images. This work takes a pioneering effort to study INRs for SAR images and finds that fine details are hard to represent. It has been shown in prior works that fine details can be easier to learn when the model weights are better initialized, which motivated us to investigate the benefits of activating the model weight before target training. The challenge of this task lies in the fact that SAR images cannot be used during the model activation stage. To this end, we propose exploiting a cross-pixel relationship of the model output, which relies on no target images. Specifically, we design a novel self-activation method by alternatively using two loss functions: a loss used to smooth out the model output, and another used for the opposite purpose. Extensive results on SAR images empirically show that our proposed method helps improve the model performance by a non-trivial margin. Full article
(This article belongs to the Section Environmental Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>Overview of the learning INR for the image signal. MLP receives pixel coordinates as input and outputs corresponding pixel values, with weight training guided by MSE loss function.</p>
Full article ">Figure 2
<p>Pipeline of the proposed method, comprising two stages: the self-activation stage and the training stage. The self-activation stage is performed before the training stage without using any information related to the target. Instead, it leverages smoothness and anti-smoothness losses to exploit the cross-pixel relationships within the model’s own output, thereby pre-activating the model.</p>
Full article ">Figure 3
<p>Visualization of the INR model SIREN’s initial output (<b>left</b>) and their corresponding gradient maps (<b>right</b>). We observe that the initial output image of SIREN, which has not been trained, is not flattened but possesses a very complex texture.</p>
Full article ">Figure 4
<p>Visualization of the NTK for the INR model SIREN with and without self-activation. After our self-activation stage, the enhanced diagonal characteristics of the model’s NTK indicate that the model is better at capturing high-frequency components.</p>
Full article ">Figure 5
<p>Qualitative evaluation of SAR image reconstruction. The first column displays the ground-truth image. Our self-activation method demonstrates enhanced capability in capturing high-frequency details.</p>
Full article ">Figure 6
<p>Qualitative evaluation of high-resolution SAR image reconstruction. The first column displays the ground-truth image, and we use a red box to highlight and zoom in on the image reconstruction details for better observation. Our self-activation method demonstrates enhanced capability in capturing high-frequency details.</p>
Full article ">Figure 7
<p>Exhibition of model performance trained with different loss functions during the training stage. Each subfigure illustrates the evolution of PSNR and SSIM metrics throughout the training stage.</p>
Full article ">Figure 8
<p>Exhibition of model performance with different numbers of self-activation stage epochs. Each subfigure illustrates the evolution of PSNR and SSIM metrics throughout the training stage.</p>
Full article ">Figure 9
<p>The model output image gradient curve over epochs during our self-activation stage. The directions denote whether the loss component is anti-smoothness or smoothness.</p>
Full article ">
19 pages, 2948 KiB  
Article
Residual-Based Implicit Neural Representation for Synthetic Aperture Radar Images
by Dongshen Han and Chaoning Zhang
Remote Sens. 2024, 16(23), 4471; https://doi.org/10.3390/rs16234471 - 28 Nov 2024
Viewed by 275
Abstract
Implicit neural representations (INRs) are a new way to represent all kinds of signals ranging from 1D audio to 3D shape signals, among which 2D images are the most widely explored due to their ubiquitous presence. Image INRs utilize a neural network to [...] Read more.
Implicit neural representations (INRs) are a new way to represent all kinds of signals ranging from 1D audio to 3D shape signals, among which 2D images are the most widely explored due to their ubiquitous presence. Image INRs utilize a neural network to learn a continuous function that takes pixel coordinates as input and outputs the corresponding pixel values. The continuous representation of synthetic aperture radar (SAR) images using INRs has not yet been explored. Existing INR frameworks developed on natural images show reasonable performance, but this performance suffers when capturing fine details. This can be attributed to INR’s prioritization of learning inter-pixel relationships, which harms intra-pixel mapping in those regions that require fine detail. To address this, we decompose the target image into an artificial uniform noise component (intra-pixel mapping) and a residual image (inter-pixel relationships). Rather than directly learning the INRs for the target image, we propose a noise-first residual learning (NRL) method. The NRL first learns the uniform noise component, then gradually incorporates the residual into the optimization target using a sine-adjusted incrementation scheme as training progresses. Given that some SAR images inherently contain significant noise, which can facilitate learning the intra-pixel independent mapping, we propose a gradient-based dataset separation method. This method distinguishes between clean and noisy images, allowing the model to learn directly from the noisy images. Extensive experimental results show that our method achieves competitive performance, indicating that learning the intra-pixel independent mapping first, followed by the inter-pixel relationship, can enhance model performance in learning INR for SAR images. Full article
Show Figures

Figure 1

Figure 1
<p>Framework for learning INR for image signals. An MLP-based network maps pixel coordinates to pixel intensity values, then is optimized by minimizing the MSE loss between the target pixel intensity values and their predicted ones.</p>
Full article ">Figure 2
<p>Plot of the periodic activation function in SIREN.</p>
Full article ">Figure 3
<p>Overview of our proposed method. The gradient-based dataset separation method distinguishes between noisy images and clean images; for clean images, we adopt NRL to learn their INRs, while noisy images are learned directly without additional operations.</p>
Full article ">Figure 4
<p>Plot of the curve of <math display="inline"><semantics> <msup> <mi>k</mi> <mi>γ</mi> </msup> </semantics></math> versus <math display="inline"><semantics> <mstyle scriptlevel="0" displaystyle="true"> <mfrac> <mi>e</mi> <mi>E</mi> </mfrac> </mstyle> </semantics></math> as <span class="html-italic">e</span> increases during the training process.</p>
Full article ">Figure 5
<p>Visualization of the representation results on both clean and noisy images. The ground truth image is presented in the first column for reference. For each of the SIREN and NRL-SIREN methods, the top row displays the represented full image and the bottom row shows a detailed portion of the image.</p>
Full article ">Figure 6
<p>Example of a clean image and a noisy image, along with their respective gradient maps.</p>
Full article ">Figure 7
<p>Illustration showing the use of a threshold <math display="inline"><semantics> <mi>α</mi> </semantics></math> to distinguish clean and noisy images.</p>
Full article ">Figure 8
<p>Qualitative comparison of SAR image representation results. Our NRL method demonstrates competitive performance in capturing fine details.</p>
Full article ">Figure 9
<p>Example results of image representation produced by various methods. The ground truth image is presented in the left-most column for reference. For each method, the top row displays the reconstructed full image, while the bottom row shows a detailed portion of the image.</p>
Full article ">Figure 10
<p>Plots the curve of <math display="inline"><semantics> <msup> <mi>k</mi> <mi>γ</mi> </msup> </semantics></math> versus <math display="inline"><semantics> <mstyle scriptlevel="0" displaystyle="true"> <mfrac> <mi>e</mi> <mi>E</mi> </mfrac> </mstyle> </semantics></math> for various <math display="inline"><semantics> <mi>γ</mi> </semantics></math> as <span class="html-italic">e</span> increases during the training process.</p>
Full article ">Figure 11
<p>Performance analysis of models trained on the target image with varying numbers of epochs. Each subfigure illustrates the progression of the PSNR and SSIM metrics during training.</p>
Full article ">Figure 12
<p>Qualitative comparison of error maps. SIRAN fails to capture fine details such as the white regions, while our proposed method helps to mitigate this problem.</p>
Full article ">
25 pages, 41258 KiB  
Article
The Deformation Monitoring Capability of Fucheng-1 Time-Series InSAR
by Zhouhang Wu, Wenjun Zhang, Jialun Cai, Hongyao Xiang, Jing Fan and Xiaomeng Wang
Sensors 2024, 24(23), 7604; https://doi.org/10.3390/s24237604 - 28 Nov 2024
Viewed by 285
Abstract
The Fucheng-1 (FC-1) satellite has successfully transitioned from its initial operational phase and is now undergoing a detailed performance assessment for time-series deformation monitoring. This study evaluates the surface deformation monitoring capabilities of the newly launched FC-1 satellite using the interferometric synthetic aperture [...] Read more.
The Fucheng-1 (FC-1) satellite has successfully transitioned from its initial operational phase and is now undergoing a detailed performance assessment for time-series deformation monitoring. This study evaluates the surface deformation monitoring capabilities of the newly launched FC-1 satellite using the interferometric synthetic aperture radar (InSAR) technique, particularly in urban applications. By analyzing the observation data from 20 FC-1 scenes and 20 Sentinel-1 scenes, deformation velocity maps of a university in Mianyang city were obtained using persistent scatterer interferometry (PSI) and distributed scatterer interferometry (DSI) techniques. The results show that thanks to the high resolution of 3 × 3 m of the FC-1 satellite, significantly more PS points and DS points were detected than those detected by Sentinel-1, by 13.4 times and 17.9 times, respectively. The distribution of the major deformation areas detected by both satellites in the velocity maps is generally consistent. FC-1 performs better than Sentinel-1 in monitoring densely structured and vegetation-covered areas. Its deformation monitoring capability at the millimeter level was further validated through comparison with leveling measurements, with average errors and root mean square errors of 1.761 mm and 2.172 mm, respectively. Its high-resolution and high-precision interferometry capabilities make it particularly promising in the commercial remote sensing market. Full article
(This article belongs to the Special Issue Recent Advances in Synthetic Aperture Radar (SAR) Remote Sensing)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Coverage areas of Sentinel-1 (purple) and FC-1 (brown), study area location marked by a five-pointed star, and COPDEM topographic map. (<b>b</b>) Google Maps image of the study area.</p>
Full article ">Figure 2
<p>Flow chart of DSI and PSI.</p>
Full article ">Figure 3
<p>(<b>a</b>) Spatio-temporal baseline map of FC-1 single master image. (<b>b</b>) Spatio-temporal baseline map of Sentinel-1 single master image.</p>
Full article ">Figure 4
<p>(<b>a</b>,<b>b</b>) Vertical deformation velocity maps from FC-1 using the DSI and PSI methods. (<b>c</b>,<b>d</b>) Vertical deformation velocity maps from Sentinel-1 using the DSI and PSI methods. (<b>e</b>) Drone orthophoto of the reference point.</p>
Full article ">Figure 5
<p>(<b>a</b>,<b>b</b>) Histograms of deformation velocity from FC-1 using the DSI and PSI methods. (<b>c</b>,<b>d</b>) Histograms of deformation velocity from Sentinel-1 using the DSI and PSI methods.</p>
Full article ">Figure 6
<p>(<b>a</b>) Schematic diagram of the research area on Google Earth. (<b>b</b>,<b>c</b>) Deformation rate maps of region R1 obtained by FC-1 and Sentinel-1 using the PSI method, with a drone image as the base map. (<b>d</b>–<b>g</b>) Deformation rate maps of regions R2 and R3 obtained by FC-1 and Sentinel-1 using the PSI method, with Google Earth as the base map. (<b>h</b>–<b>k</b>) Deformation rate maps of regions R4 and R5 obtained by FC-1 and Sentinel-1 using the DSI method, with Google Earth or a drone image as the base map.</p>
Full article ">Figure 7
<p>(<b>a</b>) Deformation velocity points obtained by FC-1 using the PSI method overlaid onto a drone image. (<b>b</b>) Deformation velocity points obtained by Sentinel-1 using the PSI method overlaid onto a drone image.</p>
Full article ">Figure 8
<p>(<b>a</b>,<b>d</b>) Deformation velocity maps from FC-1 and Sentinel-1 using the PSI method, with schematic maps of ZZ1 and ZZ2 locations. (<b>b</b>,<b>c</b>) PS deformation points from FC-1 overlaid onto drone oblique images of ZZ1 and ZZ2. (<b>e</b>,<b>f</b>) PS deformation points from Sentinel-1 overlaid onto drone oblique images of ZZ1 and ZZ2.</p>
Full article ">Figure 9
<p>(<b>a</b>,<b>b</b>) Deformation velocity points obtained by FC-1 using DSI and PSI methods overlaid onto Google imagery. (<b>c</b>,<b>d</b>) Deformation velocity points obtained by Sentinel-1 using DSI and PSI methods overlaid onto Google imagery.</p>
Full article ">Figure 10
<p>Diagram of road profile location.</p>
Full article ">Figure 11
<p>(<b>a</b>,<b>b</b>) Deformation velocity profile of FC-1 under the DSI and PSI methods. (<b>c</b>,<b>d</b>) Deformation velocity profile of Sentinel-1 under the DSI and PSI methods.</p>
Full article ">Figure 12
<p>(<b>a</b>,<b>b</b>) Deformation rate profiles of Sentinel-1 and FC-1 under the DSI method. (<b>c</b>) Diagram of position of vegetation section line. (<b>d</b>) UAV 3D model of vegetation area.</p>
Full article ">Figure 13
<p>(<b>a</b>,<b>b</b>) Coherence histograms and average coherence values for the PSI method with FC-1 and Sentinel-1. (<b>c</b>,<b>d</b>) Coherence histograms and average coherence values for the DSI method with FC-1 and Sentinel-1.</p>
Full article ">Figure 14
<p>(<b>a</b>,<b>b</b>) Standard deviation maps of deformation velocity for Sentinel-1 using PSI and DSI methods. (<b>c</b>,<b>d</b>) Standard deviation maps of deformation velocity for FC-1 using PSI and DSI methods.</p>
Full article ">Figure 15
<p>(<b>a</b>) Diagram of locations of four regions A, B, C and D. (<b>b</b>–<b>e</b>) Time-series settlement maps of FC-1 and Sentinel-1 using the DSI and PSI methods in four regions.</p>
Full article ">Figure 16
<p>(<b>a</b>–<b>d</b>) Spearman’s correlation matrix heatmaps of the time-series settlement amounts obtained by FC-1 and Sentinel-1 using the DSI and PSI methods in four regions.</p>
Full article ">Figure 17
<p>(<b>a</b>–<b>d</b>) Pearson’s correlation matrix plots of the time-series subsidence values between FC-1 and Sentinel-1 using the DSI and PSI methods in four regions.</p>
Full article ">Figure 18
<p>Illustrative Google Earth map showing the locations of level points.</p>
Full article ">Figure 19
<p>The subsidence measured by FC-1 using the DSI method compared to the subsidence measured by leveling.</p>
Full article ">Figure 20
<p>The subsidence measured by FC-1 and Sentinel-1 using the DSI method compared to the subsidence measured by leveling.</p>
Full article ">
19 pages, 8144 KiB  
Article
Thermal Optimization Design for a Small Flat-Panel Synthetic Aperture Radar Satellite
by Tian Bai, Yuanbo Zhang, Lin Kong, Hongrui Ao, Jisong Yu and Lei Zhang
Aerospace 2024, 11(12), 982; https://doi.org/10.3390/aerospace11120982 - 27 Nov 2024
Viewed by 312
Abstract
This article introduces a small microwave remote sensing satellite weighing 310 kg, operating in low earth orbit (LEO). It is equipped with an X-band synthetic aperture radar (SAR) antenna, capable of a maximum imaging resolution of 0.6 m. To achieve the objectives of [...] Read more.
This article introduces a small microwave remote sensing satellite weighing 310 kg, operating in low earth orbit (LEO). It is equipped with an X-band synthetic aperture radar (SAR) antenna, capable of a maximum imaging resolution of 0.6 m. To achieve the objectives of lower cost, reduced weight, minimized power consumption, and enhanced temperature stability, an optimized thermal design method tailored for satellites has been developed, with a particular focus on SAR antennas. The thermal control method of the antenna is closely integrated with structural design, simplifying the thermal design and its assembly process, reducing the resource consumption of thermal control systems. The distribution of thermal interface material (TIM) in the antenna assembly has been carefully calculated, achieving a zero-consumption thermal design for the SAR antenna. And the temperature difference of the entire antennas when powered on and powered off would not exceed 17 °C, meeting the specification requirements. In addition, to ensure the accuracy of antenna pointing, the support plate of antennas requires stable temperature. The layout of the heaters on the board has been optimized, reducing the use of heaters by 30% while ensuring that the temperature variation of the support board remains within 5 °C. Then, an on-orbit thermal simulation analysis of the satellite was conducted to refine the design and verification. Finally, the thermal test of the SAR satellite under vacuum conditions was conducted, involving operating the high-power antenna, verifying that the peak temperature of T/RM is below 29 °C, the temperature fluctuation amplitude during a single imaging task is 10 °C, and the lowest temperature point of the support plate is 16 °C. The results of the thermal simulation and test are highly consistent, verifying the correctness and effectiveness of the thermal design. Full article
(This article belongs to the Section Astronautics & Space Science)
Show Figures

Figure 1

Figure 1
<p>Structural layout of SAR01A: (<b>a</b>) overview of in-orbit status, (<b>b</b>) view from −Y side (PS: propulsion system; DTCU: data transmission control unit; DSU: data storage unit; ACU: antenna control unit; DCU: digital control unit; PCDU: power conditioning and distribution unit; RW: reaction wheel; NRU: navigation receiving unit; MCU: measurement and control unit).</p>
Full article ">Figure 2
<p>Heat transfer pathway. In which dotted arrows stands for the direction of heat transfer.</p>
Full article ">Figure 3
<p>Temperature control strategy for support plate.</p>
Full article ">Figure 4
<p>Thermal design strategy for antennas.</p>
Full article ">Figure 5
<p>The orbit schematic (from the perspective of the sun).</p>
Full article ">Figure 6
<p>Thermal analysis model: (<b>a</b>) antenna; (<b>b</b>) platform.</p>
Full article ">Figure 6 Cont.
<p>Thermal analysis model: (<b>a</b>) antenna; (<b>b</b>) platform.</p>
Full article ">Figure 7
<p>The simulated temperature of the antenna: (<b>a</b>) T/RM, (<b>b</b>) PM, (<b>c</b>) WPCU, and (<b>d</b>) antenna mounting plate.</p>
Full article ">Figure 7 Cont.
<p>The simulated temperature of the antenna: (<b>a</b>) T/RM, (<b>b</b>) PM, (<b>c</b>) WPCU, and (<b>d</b>) antenna mounting plate.</p>
Full article ">Figure 8
<p>The simulated temperature for the satellite: (<b>a</b>) support plate; (<b>b</b>) platform component.</p>
Full article ">Figure 8 Cont.
<p>The simulated temperature for the satellite: (<b>a</b>) support plate; (<b>b</b>) platform component.</p>
Full article ">Figure 9
<p>Heat flux simulation system with absorbing material.</p>
Full article ">Figure 10
<p>T/RMs’ temperature during TVAC testing.</p>
Full article ">Figure 11
<p>Temperature curve of the support plate: (<b>a</b>) cold case; (<b>b</b>) hot case.</p>
Full article ">
17 pages, 9078 KiB  
Article
Mapping Surface Deformation in Rwanda and Neighboring Areas Using SBAS-InSAR
by Adrien Mugabushaka, Zhenhong Li, Xuesong Zhang, Chuang Song, Bingquan Han, Bo Chen, Zhenjiang Liu and Yi Chen
Remote Sens. 2024, 16(23), 4456; https://doi.org/10.3390/rs16234456 - 27 Nov 2024
Viewed by 474
Abstract
Surface deformation poses significant risks to urban infrastructure, agriculture, and the environment in many regions worldwide, including Rwanda and the neighboring areas. This study focuses on surface deformation mapping and time series analysis in Rwanda and the neighboring areas from 2 July 2016 [...] Read more.
Surface deformation poses significant risks to urban infrastructure, agriculture, and the environment in many regions worldwide, including Rwanda and the neighboring areas. This study focuses on surface deformation mapping and time series analysis in Rwanda and the neighboring areas from 2 July 2016 to 8 June 2023 using the Small Baseline Subset Interferometric Synthetic Aperture Radar (SBAS-InSAR). The influence of atmospheric delay error is effectively reduced by integrating the Generic Atmospheric Correction Online Service (GACOS), which provides precise atmospheric delay maps. Then the SBAS-InSAR method is utilized to generate surface deformation maps and displacement time series across the region. The results of this study indicated that the maximum deformation rate was −0.11 m/yr (subsidence) and +0.13 m/yr (uplift). Through time series analysis, we quantified subsidence and uplift areas and identified key drivers of surface deformation. Since subsidence or uplift varies across the region, we have summarized the different deformation patterns and briefly analyzed the factors that may lead to deformation. Finally, this study underscores the importance of SBAS-InSAR for tracking surface deformation in Rwanda and the neighboring areas, which offers valuable perspectives for sustainable land utilization strategizing and risk mitigation. Full article
Show Figures

Figure 1

Figure 1
<p>Study area. (<b>a</b>) Map of Rwanda, including four provinces and the capital city, with provincial boundaries indicated by black lines. (<b>b</b>) Map of Africa, country boundaries indicated by blue lines.</p>
Full article ">Figure 2
<p>Spatial and temporal baselines of interferograms. Blue triangles represent Sentinel-1 SAR image acquisitions.</p>
Full article ">Figure 3
<p>Workflow of SBAS-InSAR with GACOS.</p>
Full article ">Figure 4
<p>Examples of interferograms with GACOS correction. The results before (<b>left</b>) and after correction (<b>right</b>) using GACOS were plotted using the same color and Std is the standard deviation in radian (rad).</p>
Full article ">Figure 5
<p>Linear mean velocity in line of sight (LOS) direction. The yellow pentagram represents the selected point, the red line represents the location of the known faults, and the blue and red lines in the rate chart represent uplift and subsidence, respectively. Source of fault data: <a href="https://blogs.openquake.org/hazard/global-active-fault-viewer/" target="_blank">https://blogs.openquake.org/hazard/global-active-fault-viewer/</a> (accessed on 20 October 2023). (<b>a</b>–<b>d</b>) correspond to the spatial extent outlined in the left map.</p>
Full article ">Figure 6
<p>Cumulative displacement maps from 2016 to 2023 (<b>a</b>–<b>h</b>).</p>
Full article ">Figure 7
<p>Deformation time series of different critical points. The serial numbers of critical points are labelled above each subplot. (<b>A</b>) Uplift of t he plains area along the southern shore of Lake Edward. (<b>B</b>) Subsidence in the plain’s region near Mbarara. (<b>C</b>) Deformation near Nyiragongo Volcano, (<b>D</b>) Uplift in the capital, Kigali City. (<b>E</b>) Points with subsidence trends. (<b>F</b>) Points with fluctuating deformation. (<b>G</b>) Point D31 with uplift deformation. (<b>H</b>) Point D33 with anomalous deformation time series.</p>
Full article ">Figure 8
<p>Land deformation velocity in the western part of Rwanda, including Rubavu and Nyabihu districts (south of the volcano).</p>
Full article ">Figure 9
<p>(<b>a</b>) Epicenters of earthquakes with magnitude of 4.5+ and (<b>b</b>) interferogram of the volcano-tectonic event (20210513-2021052).</p>
Full article ">Figure 10
<p>Details of the time series of points located south and east of Nyiragongo volcano. The green dotted line indicates the 2021 eruption of Nyiragongo.</p>
Full article ">Figure 11
<p>Cracks of some buildings in the Rubavu district of Western part of Rwanda due to 2021 volcano-tectonic event. The photos were taken near the key point D20. The red arrows indicate the cracks on the walls. (<b>a</b>–<b>d</b>) show four different examples of wall cracks.</p>
Full article ">Figure 12
<p>Rwanda land use/land cover change.</p>
Full article ">Figure 13
<p>(<b>a</b>) Relation between annual rainfall and land deformation and (<b>b</b>) velocity map of Kigali city. Data were downloaded from the Rwanda Meteorological Agency (<a href="https://www.meteorwanda.gov.rw/index.php?id=2" target="_blank">https://www.meteorwanda.gov.rw/index.php?id=2</a> (accessed on 21 October 2023).</p>
Full article ">
31 pages, 2697 KiB  
Article
Single-Scene SAR Image Data Augmentation Based on SBR and GAN for Target Recognition
by Shangchen Feng, Xikai Fu, Yanlin Feng and Xiaolei Lv
Remote Sens. 2024, 16(23), 4427; https://doi.org/10.3390/rs16234427 - 26 Nov 2024
Viewed by 260
Abstract
High-performance neural networks for synthetic aperture radar (SAR) automatic target recognition (ATR) often encounter the challenge of data scarcity. The lack of sufficient labeled SAR image datasets leads to the consideration of using simulated data to supplement the dataset. On the one hand, [...] Read more.
High-performance neural networks for synthetic aperture radar (SAR) automatic target recognition (ATR) often encounter the challenge of data scarcity. The lack of sufficient labeled SAR image datasets leads to the consideration of using simulated data to supplement the dataset. On the one hand, electromagnetic computation simulations provide high amplitude accuracy but are inefficient for large-scale datasets due to their complex computations and physical models. On the other hand, ray tracing simulations offer high geometric accuracy and computational efficiency but struggle with low amplitude correctness, hindering accurate numerical feature extraction. Furthermore, the emergence of generative adversarial networks (GANs) provides a way to generate simulated datasets, trying to balance computational efficiency with image quality. Nevertheless, the simulated SAR images generated based on random noise lack constraints, and it is also difficult to generate images that exceed the parameter conditions of the real image’s training set. Hence, it is essential to integrate physics-based simulation techniques into GANs to enhance the generalization ability of the imaging parameters. In this paper, we present the SingleScene-SAR Simulator, an efficient framework for SAR image simulation that operates under limited real SAR data. This simulator integrates rasterized shooting and bouncing rays (SBR) with cycle GAN, effectively achieving both amplitude correctness and geometric accuracy. The simulated images are appropriate for augmenting datasets in target recognition networks. Firstly, the SingleScene-SAR Simulator employs a rasterized SBR algorithm to generate radar cross section (RCS) images of target models. Secondly, a specific training pattern for cycle GAN is established to translate noisy RCS images into simulated SAR images that closely resemble real ones. Finally, these simulated images are utilized for data augmentation. Experimental results based on the constructed dataset show that with only one scene SAR image containing 30 target chips, the SingleScene-SAR Simulator can efficiently produce simulated SAR images that exhibit high similarity in both spatial and statistical distributions compared with real images. By employing simulated SAR images for data augmentation, the accuracy of target recognition networks can be consistently and significantly enhanced. Full article
Show Figures

Figure 1

Figure 1
<p>Shooting and bouncing rays.</p>
Full article ">Figure 2
<p>(<b>a</b>) Snell’s Law. (<b>b</b>) Ray tube reflection on the triangular facet.</p>
Full article ">Figure 3
<p>The network architecture of cycle GAN.</p>
Full article ">Figure 4
<p>Framework of the SingleScene–SAR simulator. (<b>a</b>) The component of RCS image simulation. (<b>b</b>) The component of SAR image translation.</p>
Full article ">Figure 5
<p>(<b>a</b>) Spaceborne SAR platform motion model. (<b>b</b>) The grid sampling of the target model and the shape vectors of the ray pool.</p>
Full article ">Figure 6
<p>Aircraft recognition network architecture. (<b>a</b>) ConvNet. (<b>b</b>) VGGNet.</p>
Full article ">Figure 7
<p>Parameter of SAR images from the Aircraft-VariedSAR dataset.</p>
Full article ">Figure 8
<p>Aircraft models.</p>
Full article ">Figure 9
<p>The real SAR image (<b>a</b>) and its 90° rotated counterpart (<b>b</b>). The SSIM of these two images is 0.1921.</p>
Full article ">Figure 10
<p>Real SAR images, simulated RCS images, and simulated SAR images generated from the RCS2SAR-30/40/50 datasets. The 1st and 2nd rows display real SAR images; the 3rd and 4th rows showcase simulated RCS images. The final 6 rows consist of simulated SAR images derived from the RCS2SAR-30/40/50 datasets, with dataset names indicated on the left side. In addition to the real SAR images, corresponding CAD models and incident angles for each image are labeled above each column. The angle values located in the bottom right corner of the images denote the rotation angle of the CAD model around the z-axis.</p>
Full article ">Figure 11
<p>Normalized histogram results.</p>
Full article ">Figure 12
<p>F1-score curves. The experiments are organized from top to bottom according to the simulated datasets RCS2SAR-30, RCS2SAR-40, and RCS2SAR-50. From <b>left</b> to <b>right</b>, the experiments utilize ConvNet and VGGNet architectures with double-sized clutters (D) and triple-sized clutters (T), respectively.</p>
Full article ">Figure 13
<p>Simulated images derived from the ablation experiments. The 1st row shows images from “SBR and Noise Estimator only”; the remaining 3 rows show images from “SBR and cycle GAN only”, with the names of the corresponding simulated datasets of baseline labeled on the left. The angle values in the bottom right corner of each image represent the clockwise rotation angle of the CAD model around the z-axis.</p>
Full article ">
18 pages, 2075 KiB  
Article
Multiple-Input Multiple-Output Synthetic Aperture Radar Waveform and Filter Design in the Presence of Uncertain Interference Environment
by Ke Xu, Guohao Sun, Yuandong Ji, Zhiquan Ding and Wenhao Chen
Remote Sens. 2024, 16(23), 4413; https://doi.org/10.3390/rs16234413 - 25 Nov 2024
Viewed by 343
Abstract
Multiple-input multiple-output synthetic aperture radar (MIMO-SAR) anti-jamming waveform design relies on accurate prior information about the interference. However, it is difficult to obtain accurate prior knowledge about uncertain intermittent sampling repeater jamming (ISRJ), leading to a severe decline in the detection performance of [...] Read more.
Multiple-input multiple-output synthetic aperture radar (MIMO-SAR) anti-jamming waveform design relies on accurate prior information about the interference. However, it is difficult to obtain accurate prior knowledge about uncertain intermittent sampling repeater jamming (ISRJ), leading to a severe decline in the detection performance of MIMO-SAR systems. Therefore, this article studies the robust joint design problem of MIMO radar transmit waveform and filter against uncertain ISRJ. We characterize two categories of uncertain interference, including sample length uncertainty and sample-time uncertainty, modeled as Gaussian distribution in different range bins. Based on the uncertain interference model, we formulate the maximizing SINR as a figure of merit, which is a non-convex quadratic optimization problem under specific waveform constraints. Based on the alternating direction method of multipliers (ADMM) framework, a novel joint design algorithm of waveform and filter is proposed. In order to improve the convergence performance of ADMM, the difference in convex functions (DC) programming is applied to the ADMM iterations framework to solve the problem of waveform energy inequality constraint. Finally, numerical results demonstrate the effectiveness and robustness of the proposed method, compared to the existing methods that utilize deterministic interference models in the uncertain ISRJ environment. Moreover, the spaceborne SAR real scene imaging simulations are conducted to evaluate the anti-ISRJ performance. Full article
Show Figures

Figure 1

Figure 1
<p>Uncertain store-and-forward schematic. (<b>a</b>) sample-length; (<b>b</b>) sample-time; (<b>c</b>) sample-length and sample-time.</p>
Full article ">Figure 2
<p>Schematic representation of the uncertainty in the location of the source of interference.</p>
Full article ">Figure 3
<p>Iterative output SINR for ISRJ under sample-length uncertainty environment. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 4
<p>Comparison of the robustness of different waveforms under sample−length uncertainty environment. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low−energy waveform design condition.</p>
Full article ">Figure 5
<p>Estimation error of different waveforms regarding DOA localization. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 6
<p>Interference spectra under uncertain sample-length environment. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 7
<p>Iterative output SINR for ISRJ under sample-time uncertainty environment. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 8
<p>Comparison of the robustness of different waveform under sampling time uncertainties. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 9
<p>Estimation error of different waveform regarding DOA localization. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 10
<p>Interference spectra under uncertainty of sample time. (<b>a</b>) unit energy waveform design condition; (<b>b</b>) low-energy waveform design condition.</p>
Full article ">Figure 11
<p>Imaging results in the presence of uncertain ISRJ. (<b>a</b>) image without uncertain ISRJ suppression; (<b>b</b>) image obtained using waveform of the proposed method.</p>
Full article ">Figure 12
<p>Imaging results in the presence of uncertain ISRJ. (<b>a</b>) image without uncertain ISRJ suppression; (<b>b</b>) image obtained using waveform of the proposed method.</p>
Full article ">
17 pages, 18667 KiB  
Article
Enhanced Wide-Area Glacier Velocity Monitoring in Svalbard via Synthetic Aperture Radar Offset Tracking Noise Suppression
by Honglei Yang, Songxue Zhao, Zeping Wang, Ao Yan and Zhenhan Shi
Appl. Sci. 2024, 14(23), 10834; https://doi.org/10.3390/app142310834 - 22 Nov 2024
Viewed by 381
Abstract
Glacier movement is an important indicator of climate change, reflecting the quality and state changes in glacier migration and mass balance in the context of global warming. Although accurately estimating glacier surface flow velocity is crucial for various applications, achieving this is challenging [...] Read more.
Glacier movement is an important indicator of climate change, reflecting the quality and state changes in glacier migration and mass balance in the context of global warming. Although accurately estimating glacier surface flow velocity is crucial for various applications, achieving this is challenging due to factors such as low temporal correlation and high noise effects. This paper presents the pixel offset tracking (POT) technology based on Synthetic Aperture Radar (SAR) data for glacier velocity monitoring, with enhanced cross-correlation matching window and noise suppression approaches. In particular, a noise suppression optimization method and a matching window optimization index suitable for wide-area glacier velocity monitoring are proposed. The inter-annual wide-area two-dimensional plane flow velocity of glaciers in the Svalbard archipelago was obtained by using a total of seven Sentinel-1 data sets from two orbits covering the entire Svalbard archipelago in 2021. The results indicate that 25 large glaciers in Svalbard destabilized in 2021, with a peak flow velocity of 6.18 m/day. At the same time, the influence of climate, topography, and other factors on glacier surface velocity is discussed. The wide-area glacier velocity monitoring method and its application demonstrated in this paper will serve as a valuable reference for studying glacier migration in the Arctic Svalbard archipelago and for other large-scale wide-area deformation monitoring efforts. Full article
(This article belongs to the Special Issue Latest Advances in Radar Remote Sensing Technologies)
Show Figures

Figure 1

Figure 1
<p>(<b>a</b>) Svalbard geographical location and Sentinel-1 image coverage used in research. (<b>b</b>) Study area DSM.</p>
Full article ">Figure 2
<p>Sentinel-1 image intensity of Monacobreen glacier in Svalbard selected from the pre-experiment.</p>
Full article ">Figure 3
<p>Results of the first (<b>a-1</b>–<b>a-3</b>,<b>b-1</b>–<b>b-3</b>) and second (<b>c-1</b>–<b>c-3</b>,<b>d-1</b>–<b>d-3</b>) image pairs along the range and azimuth under different cross-correlation windows in the pre-experiment.</p>
Full article ">Figure 3 Cont.
<p>Results of the first (<b>a-1</b>–<b>a-3</b>,<b>b-1</b>–<b>b-3</b>) and second (<b>c-1</b>–<b>c-3</b>,<b>d-1</b>–<b>d-3</b>) image pairs along the range and azimuth under different cross-correlation windows in the pre-experiment.</p>
Full article ">Figure 4
<p>(<b>a</b>) Flow velocity of the Svalbard wide-area glacier. (<b>b</b>) Corresponding slope information of Svalbard wide-area glacier.</p>
Full article ">Figure 5
<p>The cumulative displacement of the Monacobreen glacier under different time baselines in 2021.</p>
Full article ">Figure 6
<p>(<b>a</b>) The corresponding changes in time-series flow velocity on the glacier surface with temperature. (<b>b</b>) The corresponding changes in time-series flow velocity on the glacier surface with accumulated precipitation over time.</p>
Full article ">Figure 7
<p>(<b>a</b>) Schematic diagram of the glacier profile line. (<b>b</b>) Longitudinal profile of flow velocity and elevation changes. (<b>c</b>) Transverse profile of flow velocity and elevation changes.</p>
Full article ">Figure 8
<p>The flow velocity change in each sample when the elevation rises by one meter.</p>
Full article ">Figure 9
<p>Schematic of the wedge pattern of glacier flow.</p>
Full article ">
27 pages, 25812 KiB  
Article
Forecasting Flood Inundation in U.S. Flood-Prone Regions Through a Data-Driven Approach (FIER): Using VIIRS Water Fractions and the National Water Model
by Amirhossein Rostami, Chi-Hung Chang, Hyongki Lee, Hung-Hsien Wan, Tien Le Thuy Du, Kel N. Markert, Gustavious P. Williams, E. James Nelson, Sanmei Li, William Straka III, Sean Helfrich and Angelica L. Gutierrez
Remote Sens. 2024, 16(23), 4357; https://doi.org/10.3390/rs16234357 - 22 Nov 2024
Viewed by 494
Abstract
Floods, one of the costliest, and most frequent hazards, are expected to worsen in the U.S. due to climate change. The real-time forecasting of flood inundations is extremely important for proactive decision-making to reduce damage. However, traditional forecasting methods face challenges in terms [...] Read more.
Floods, one of the costliest, and most frequent hazards, are expected to worsen in the U.S. due to climate change. The real-time forecasting of flood inundations is extremely important for proactive decision-making to reduce damage. However, traditional forecasting methods face challenges in terms of implementation and scalability due to computational burdens and data availability issues. Current forecasting services in the U.S. largely rely on hydrodynamic modeling, limited to river reaches near in situ gauges and requiring extensive data for model setup and calibration. Here, we have successfully adapted the Forecasting Inundation Extents using REOF (FIER) analysis framework to produce forecasted water fraction maps in two U.S. flood-prone regions, specifically the Red River of the North Basin and the Upper Mississippi Alluvial Plain, utilizing Visible Infrared Imaging Radiometer Suite (VIIRS) optical imagery and the National Water Model. Comparing against historical VIIRS imagery for the same dates, FIER 1- to 8-day medium-range pseudo-forecasts show that about 70–80% of pixels exhibit absolute errors of less than 30%. Although originally developed utilizing Synthetic Aperture Radar (SAR) images, this study demonstrated FIER’s versatility and effectiveness in flood forecasting by demonstrating its successful adaptation with optical VIIRS imagery which provides daily water fraction product, offering more historical observations to be used as inputs for FIER during peak flood times, particularly in regions where flooding commonly happens in a short period rather than following a broad seasonal pattern. Full article
Show Figures

Figure 1

Figure 1
<p>(<b>top</b>) The USGS in situ streamflow data (blue line, cumecs: m<sup>3</sup>/second) from 2017 to 2020 at gauges located in (<b>a</b>) Drayton, North Dakota, along the Red River of the North mainstem, and (<b>b</b>) New Madrid, Missouri, along the Mississippi River mainstem. The green triangles mark all the epochs when Sentinel-1 images were acquired, while the orange dots mark the epochs of the VIIRS images used in this study. (<b>bottom</b>) The corresponding amount of data with less than 5% cloud coverage within each of the 10% USGS in situ streamflow percentile groups.</p>
Full article ">Figure 2
<p>(<b>Left column</b>) JRC historical maximum inundation extents and permanent water from 1984 to 2022 [<a href="#B42-remotesensing-16-04357" class="html-bibr">42</a>], and (<b>right column</b>) the USGS NLCD 2021 cultivated croplands [<a href="#B44-remotesensing-16-04357" class="html-bibr">44</a>] in (<b>a</b>) RRNB and (<b>b</b>) UMAP. The white dots show the locations of the USGS in situ gauges used in this study.</p>
Full article ">Figure 3
<p>Pie charts of the top five classes in the USDA CDL for 2021 and 2022 in the (<b>a</b>) RRNB and (<b>b</b>) UMAP, showing the most dominant crops are spring wheat and soybeans, respectively.</p>
Full article ">Figure 4
<p>Flowchart of the FIER process largely consists of framework construction and forecasting. Dashed line arrow indicates the synthesis of RSMs and forecasted RTPCs.</p>
Full article ">Figure 5
<p>Flowchart (<b>left</b>) and schematic view (<b>right</b>) of the quantile mapping process employed to correct the biases in FIER water fraction forecasts. The blue boxes in the flowchart represent historical water fraction data (FIER-synthesized and VIIRS-observed) and their respective CDFs. The red boxes in the flowchart represent forecasted water fraction data and the corresponding extracted quantiles.</p>
Full article ">Figure 6
<p>The extracted streamflow-related (<b>a</b>) RSMs, (<b>b</b>) RTPCs along with USGS in situ streamflow data, and (<b>c</b>) neural network regression models for FIER water fraction forecasting in the RRNB.</p>
Full article ">Figure 7
<p>The extracted streamflow-related (<b>a</b>) RSMs, (<b>b</b>) RTPCs along with USGS in situ streamflow data, and (<b>c</b>) neural network regression models for FIER water fraction forecasting in the UMAP.</p>
Full article ">Figure 8
<p>The extracted streamflow-related (<b>a</b>) RSMs, (<b>b</b>) RTPCs along with USGS in situ streamflow data, and (<b>c</b>) neural network regression models for FIER Sentinel-1 inundation extent forecasting in the RRNB.</p>
Full article ">Figure 9
<p>The extracted streamflow-related (<b>a</b>) RSMs, (<b>b</b>) RTPCs along with USGS in situ streamflow data, and (<b>c</b>) neural network regression models for FIER Sentinel-1 inundation extent forecasting in the UMAP.</p>
Full article ">Figure 10
<p>Cumulative percentages of pixels in different ranges of AEs in the RRNB and UMAP over (<b>a</b>) all pixels and (<b>b</b>) pixels with high water fractions (&gt;80%).</p>
Full article ">Figure 11
<p>Water fractions on the peak-flood dates in 2022 and 2023 in the RRNB: (<b>a</b>) historical observation where white pixels are clouds, (<b>b</b>) FIER pseudo-nowcast, and (<b>c</b>) 8-day FIER medium-range pseudo-forecast.</p>
Full article ">Figure 12
<p>Water fractions on the peak flood dates in 2021, 2022, and 2023 in the UMAP: (<b>a</b>) historical observation where white pixels are clouds, (<b>b</b>) FIER pseudo-nowcast, and (<b>c</b>) 8-day FIER medium-range pseudo-forecast.</p>
Full article ">Figure 13
<p>Examples of averaged FIER medium-range water fraction pseudo-forecasts over the next 1 to 8 days in the 2022 spring wheat fields in the RRNB, which could have been generated on (<b>a</b>) 2022-05-02, (<b>b</b>) 2022-05-03, or (<b>c</b>) 2022-05-04, before the peak flood on 2022-05-05 in the planting period.</p>
Full article ">Figure 14
<p>Examples of averaged FIER medium-range water fraction pseudo-forecasts over the next 1 to 8 days in the 2022 soybean fields in the UMAP, which could have been generated on (<b>a</b>) 2022-05-11, (<b>b</b>) 2022-05-12, or (<b>c</b>) 2022-05-13, before the peak flood on 2022-05-14 in the planting period.</p>
Full article ">
21 pages, 2229 KiB  
Article
LH-YOLO: A Lightweight and High-Precision SAR Ship Detection Model Based on the Improved YOLOv8n
by Qi Cao, Hang Chen, Shang Wang, Yongqiang Wang, Haisheng Fu, Zhenjiao Chen and Feng Liang
Remote Sens. 2024, 16(22), 4340; https://doi.org/10.3390/rs16224340 - 20 Nov 2024
Viewed by 552
Abstract
Synthetic aperture radar is widely applied to ship detection due to generating high-resolution images under diverse weather conditions and its penetration capabilities, making SAR images a valuable data source. However, detecting multi-scale ship targets in complex backgrounds leads to issues of false positives [...] Read more.
Synthetic aperture radar is widely applied to ship detection due to generating high-resolution images under diverse weather conditions and its penetration capabilities, making SAR images a valuable data source. However, detecting multi-scale ship targets in complex backgrounds leads to issues of false positives and missed detections, posing challenges for lightweight and high-precision algorithms. There is an urgent need to improve accuracy of algorithms and their deployability. This paper introduces LH-YOLO, a YOLOv8n-based, lightweight, and high-precision SAR ship detection model. We propose a lightweight backbone network, StarNet-nano, and employ element-wise multiplication to construct a lightweight feature extraction module, LFE-C2f, for the neck of LH-YOLO. Additionally, a reused and shared convolutional detection (RSCD) head is designed using a weight sharing mechanism. These enhancements significantly reduce model size and computational demands while maintaining high precision. LH-YOLO features only 1.862 M parameters, representing a 38.1% reduction compared to YOLOv8n. It exhibits a 23.8% reduction in computational load while achieving a mAP50 of 96.6% on the HRSID dataset, which is 1.4% higher than YOLOv8n. Furthermore, it demonstrates strong generalization on the SAR-Ship-Dataset with a mAP50 of 93.8%, surpassing YOLOv8n by 0.7%. LH-YOLO is well-suited for environments with limited resources, such as embedded systems and edge computing platforms. Full article
Show Figures

Figure 1

Figure 1
<p>The overall structure of YOLOv8. YOLOv8 originates from the open-source code made available by Ultralytics. “ ×2” means there are two columns of ConvModule.</p>
Full article ">Figure 2
<p>The structure of LH-YOLO.</p>
Full article ">Figure 3
<p>(<b>a</b>) The fundamental module for the StarNet-nano network. (<b>b</b>) Detailed description of the star operation.</p>
Full article ">Figure 4
<p>(<b>a</b>) The C2f module in the neck of YOLOv8n. (<b>b</b>) The proposed LFE-C2f module.</p>
Full article ">Figure 5
<p>(<b>a</b>) The framework of the decoupled head of YOLOv8n. (<b>b</b>) The framework of the RSCD head.</p>
Full article ">Figure 6
<p>Detection results comparison using the HRSID dataset. Green boxes represent ships that have been correctly detected, while blue boxes indicate incorrectly detected ships. Additionally, red boxes denote ships that were not detected at all.</p>
Full article ">Figure 7
<p>Detection results comparison using the SAR-Ship-Dataset.</p>
Full article ">Figure 8
<p>Visual comparison of detection results between LH-YOLO and four detection models using the HRSID and SAR-Ship-Dataset: (<b>a</b>) Ground Truth, (<b>b</b>) YOLOv3-Tiny. (<b>c</b>) YOLOv5. (<b>d</b>) YOLOv10n. (<b>e</b>) Proposed LH-YOLO.</p>
Full article ">
Back to TopTop