Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Monitoring Surface Water Inundation of Poyang Lake and Dongting Lake in China Using Sentinel-1 SAR Images
Next Article in Special Issue
Transmit Antenna Selection and Power Allocation for Joint Multi-Target Localization and Discrimination in MIMO Radar with Distributed Antennas under Deception Jamming
Previous Article in Journal
Integrated Validation of Coarse Remotely Sensed Evapotranspiration Products over Heterogeneous Land Surfaces
Previous Article in Special Issue
Generalized Labeled Multi-Bernoulli Multi-Target Tracking with Doppler-Only Measurements
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DU-CG-STAP Method Based on Sparse Recovery and Unsupervised Learning for Airborne Radar Clutter Suppression

Air and Missile Defense College, Air Force Engineering University, Xi’an 710051, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(14), 3472; https://doi.org/10.3390/rs14143472
Submission received: 22 June 2022 / Revised: 13 July 2022 / Accepted: 14 July 2022 / Published: 19 July 2022
(This article belongs to the Special Issue Small or Moving Target Detection with Advanced Radar System)
Graphical abstract
">
Figure 1
<p>Geometry model of airborne radar.</p> ">
Figure 2
<p>Processing framework of DU-CG-STAP.</p> ">
Figure 3
<p>Network structure of 2DMA-Net.</p> ">
Figure 4
<p>Processing framework of CycleGAN.</p> ">
Figure 5
<p>Network structures of the generator and discriminator of the CycleGAN in this study.</p> ">
Figure 6
<p>Partitioning method of the generated datasets.</p> ">
Figure 7
<p>Network training convergence results: (<b>a</b>) training loss curves of 2DMA-Net and CycleGAN and (<b>b</b>) discrimination probability curves of the discriminator G<span class="html-italic"><sub>Z</sub></span>.</p> ">
Figure 8
<p>Clutter spectrum estimation results obtained using different methods in the ideal case.</p> ">
Figure 8 Cont.
<p>Clutter spectrum estimation results obtained using different methods in the ideal case.</p> ">
Figure 9
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 1.34.</p> ">
Figure 10
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with a non-side-looking angle of 16.5°.</p> ">
Figure 11
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with the ICM as 0.5 m/s.</p> ">
Figure 11 Cont.
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with the ICM as 0.5 m/s.</p> ">
Figure 12
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with the element amplitude/phase error as 0.14/3.83°.</p> ">
Figure 13
<p>Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 0.67, the non-side-looking angle as 15.50°, the ICM as 0.24 m/s, and the array element amplitude/phase error as 0.10/4°.</p> ">
Figure 14
<p>SCNR loss curves of different methods for a target with zero spatial frequency and varying Doppler frequency; subfigures (<b>a</b>–<b>f</b>) correspond to <a href="#remotesensing-14-03472-f008" class="html-fig">Figure 8</a>, <a href="#remotesensing-14-03472-f009" class="html-fig">Figure 9</a>, <a href="#remotesensing-14-03472-f010" class="html-fig">Figure 10</a>, <a href="#remotesensing-14-03472-f011" class="html-fig">Figure 11</a>, <a href="#remotesensing-14-03472-f012" class="html-fig">Figure 12</a> and <a href="#remotesensing-14-03472-f013" class="html-fig">Figure 13</a>, respectively.</p> ">
Figure 15
<p>Computational complexities of different algorithms under different conditions: (<b>a</b>) the computational complexity varies with the iteration number and (<b>b</b>) the computational complexity varies with the array element number.</p> ">
Figure 16
<p>Low-accuracy and high-accuracy mapping results of CycleGAN with the low-accuracy clutter spectrum dataset obtained by (<b>a</b>) Fourier transform, (<b>b</b>) MVDR algorithm, (<b>c</b>) MMV-ADMM algorithm, and (<b>d</b>) the trained 2DMA-Net.</p> ">
Review Reports Versions Notes

Abstract

:
With a small number of training range cells, sparse recovery (SR)-based space–time adaptive processing (STAP) methods can help to suppress clutter and detect targets effectively for airborne radar. However, SR algorithms usually have problems of high computational complexity and parameter-setting difficulties. More importantly, non-ideal factors in practice will lead to the degraded clutter suppression performance of SR-STAP methods. Based on the idea of deep unfolding (DU), a space–time two-dimensional (2D)-decoupled SR network, namely 2DMA-Net, is constructed in this paper to achieve a fast clutter spectrum estimation without complicated parameter tuning. For 2DMA-Net, without using labeled data, a self-supervised training method based on raw radar data is implemented. Then, to filter out the interferences caused by non-ideal factors, a cycle-consistent adversarial network (CycleGAN) is used as the image enhancement process for the clutter spectrum obtained using 2DMA-Net. For CycleGAN, an unsupervised training method based on unpaired data is implemented. Finally, 2DMA-Net and CycleGAN are cascaded to achieve a fast and accurate estimation of the clutter spectrum, resulting in the DU-CG-STAP method with unsupervised learning, as demonstrated in this paper. The simulation results show that, compared to existing typical SR-STAP methods, the proposed method can simultaneously improve clutter suppression performance and reduce computational complexity.

Graphical Abstract">

Graphical Abstract

1. Introduction

By simultaneously using spatial information and temporal information, the space–time adaptive processing (STAP) method can improve the clutter suppression and moving target detection performance for airborne radar [1,2]. However, to ensure that the loss of the output signal-to-clutter-plus-noise ratio (SCNR) does not exceed 3 dB compared to the optimal case, the number of independent identically distributed (IID) training range cells required by conventional STAP methods is at least twice the system degrees of freedom (DOF) [3]. In practice, non-ideal factors, e.g., a non-uniform ground/sea environment, non-stationary clutter features, complicated platform movements, and array amplitude/phase errors, often make this condition difficult to meet [4,5,6].
To reduce the requirement of IID training range cells, dimension-reduced STAP methods, rank-reduced STAP methods, direct-data-domain STAP methods, and SR-based STAP methods have been proposed [7,8,9,10]. Among these methods, SR-STAP methods can achieve a high-resolution estimation of the clutter spectrum using a small number of IID training range cells. However, most SR algorithms, e.g., the focal under-determined system solver (FOCUSS), alternating direction method of multipliers (ADMM), and fast-converging sparse Bayesian learning (FCSBL) algorithm [11,12,13] require lots of iterations to obtain the convergent solution, leading to high computational costs, especially when the problem dimension is high. In addition, in different clutter environments, the appropriate parameter settings are also a difficult problem for SR-STAP methods. Unreasonable parameter settings will affect the convergence speed and accuracy of SR algorithms. More importantly, various non-ideal factors in practical applications will deteriorate the clutter sparsity and make the SR signal model inaccurate, resulting in significant interferences deviating from the clutter ridge in the space–time domain and thus degrading the clutter suppression performance of SR-STAP methods. These problems limit the applications of SR-STAP methods in practice [14,15,16].
Recently, the deep neural network (DNN)-based deep learning (DL) technique has been developed and applied to various fields. After proper and sufficient training, DNN can obtain a powerful nonlinear transform capacity for many data-processing or feature-mapping problems [17,18,19,20]. In addition, after offline training, DNN only needs forward propagation to complete its operations, thus enjoying a high online computing efficiency. These two properties of DNN can help to solve the above-mentioned problems of SR-STAP methods. For example, a STAP method based on convolutional DNN (CNN) is proposed in [21], which uses the nonlinear image enhancement capability of CNN to realize the high-accuracy reconstruction of the clutter spectrum from its low-accuracy counterpart. It is shown in [21] that, compared to some typical SR-STAP methods, the CNN-based STAP method can obtain a higher clutter-suppression performance with lower computational costs.
Unlike classical data-driven-only DNNs, deep unfolding (DU)-based neural networks combine the data-driven method with the model-driven method [22,23,24,25]. In DU-Net, a specific iterative algorithm (e.g., an iterative SR algorithm) with given iterations is unfolded into a DNN with the same number of layers, then the parameters involved in this algorithm are optimized by data learning. In other words, the DU-Net is constructed based on the model of an iterative algorithm. Compared to data-driven DNNs, DU-Nets have the advantage of interpretability and compared to model-driven algorithms, DU-Nets have the advantages of convergence speed and accuracy. Hence, DU-Nets also have the capability to solve the problems of SR-STAP methods. For example, the ADMM algorithm is unfolded into a DNN in [26] for the joint estimation of the clutter spectrum and array error parameters. It is shown in [26] that compared to some typical SR-STAP methods, the DU-Net-based STAP method can improve the clutter suppression performance and reduce the computing complexity.
However, although showing promising potential, DNN-based and DU-Net-based STAP methods have some essential problems that need to be solved. For the STAP methods using the nonlinear image enhancement capability of DNNs, the clutter spectrum estimation performance largely depends on the quality of the input data [27], which cannot be guaranteed using conventional spectrum estimation methods, e.g., the Fourier transform method used in [26]. For the DU-Net-based STAP methods that use the SR algorithm and the DNN method jointly, the performance will be seriously degraded when the clutter sparsity is damaged by the non-ideal practical factors. In addition, for both the DNN-based and DU-Net-based STAP methods, it is usually difficult to construct sufficient and complete input-label paired datasets for supervised training in an unknown environment.
To solve these problems, a DU-CG-STAP method with unsupervised learning is proposed in this paper, which cascades a DU-Net and a DNN. The DU-Net, named as 2DMA-Net, is used to process the raw radar data to estimate the clutter spectrum. It is constructed by unfolding a space–time 2D-decoupled SR algorithm with the multiple-measurement vector (MMV) model, i.e., the 2D-MMV-ADMM algorithm. Similar to [28], the self-supervised learning method based on raw radar data is adopted by 2DMA-Net. The DNN, named a cycle-consistent adversarial network (CycleGAN) [29], is used to process the clutter spectrum obtained by 2DMA-Net to filter out the interferences caused by non-ideal factors. It is acting as a nonlinear image enhancement process with an unsupervised training method based on an input-label unpaired dataset. By using DU-Net and DNN simultaneously, the DU-CG-STAP method can realize a fast and accurate estimation of the clutter spectrum and thus achieve a high clutter suppression and target detection performance for airborne radar.
To summarize, the main contributions of this paper are as follows:
(1)
To reduce the complexity of solving the SR-STAP model for estimating the clutter spectrum, the MMV-ADMM algorithm is space–time 2D-decoupled. To optimize the iteration parameters of 2D-MMV-ADMM, the 2DMA-Net is constructed. To train 2DMA-Net, the L1 regularization loss function and the mean squared error (MSE) loss function are combined, and thus, with only raw radar data, the self-supervised training method is implemented.
(2)
To solve the performance degradation problem of SR-STAP under non-ideal conditions, the clutter spectrum obtained using 2DMA-Net is processed using CycleGAN. The generator of CycleGAN maps the low-accuracy clutter spectrum into a high-accuracy domain to adaptively extract the clutter features and thus suppress the interferences caused by non-ideal factors. With an unpaired dataset, CycleGAN is trained based on the adversarial criterion and the cycle-consistency criterion.
(3)
To generate an accurate clutter spectrum with low complexity, 2DMA-Net and CycleGAN are cascaded to form the DU-CG-STAP method. With raw radar data and theoretical clutter spectrum as the unpaired dataset, the DU-CG-STAP is trained in an unsupervised way.
The rest of this paper is organized as follows. Section 2 establishes the signal model and briefly introduces the SR-STAP method. Section 3 introduces the processing framework, network structure, dataset construction, and training methods of DU-CG-STAP in detail. Section 4 verifies the performance and advantages of the proposed method via various simulations. Section 5 draws conclusions and discusses future work.

2. Signal Model

2.1. STAP

As shown in Figure 1, consider an airborne phased-array radar moves at a constant speed v along the y-axis at an altitude of H . The number of elements in the uniform linear array (ULA) is M and the spacing between adjacent array elements is d . The angle between the ULA and the airborne moving direction (i.e., the non-side-looking angle) is θ e . The radar transmits and receives a total of N pulses in a coherent processing interval (CPI) with a pulse repetition interval of T r .
Without considering the effect of range ambiguity, the range ring on the ground/sea surface corresponding to each range cell is supposed to consist of N c clutter patches with mutually independent scattering coefficients. Thus, the clutter-plus-noise component contained in the radar-received signal x from the range cell under test (RUT) can be expressed as
x c + x n = n = 1 N c σ c ; n v ( f c ; d , n , f c ; s , n ) α ( n ) + x n = n = 1 N c σ c ; n [ v d ( f c ; d , n ) v s ( f c ; s , n ) ] [ α d ( n ) α s ( n ) ] + x n
where x n is the noise signal, which can usually be assumed to be complex Gaussian white noise with a mean of 0 and a variance of σ n 2 , denotes the Kronecker product, σ c ; n denotes the scattering coefficient of the nth clutter patch, α d ( n ) and α s ( n ) denotes the weighting vectors related to the temporal and spatial non-ideal factors (e.g., internal clutter motion (ICM), array amplitude error, and array phase error), v d ( f c ; d , n ) and v s ( f c ; s , n ) are the steering vectors of the nth clutter patch in the time and space domain, expressed as
{ v d ( f c ; d , n ) = [ 1 , exp ( j 2 π f c ; d , n ) , , exp ( j 2 π ( N 1 ) f c ; d , n ) ] T N × 1 v s ( f c ; s , n ) = [ 1 , exp ( j 2 π f c ; s , n ) , , exp ( j 2 π ( M 1 ) f c ; s , n ) ] T M × 1
where [ ] T denotes transpose operation, f c ; d , n and f c ; s , n are the Doppler frequency and spatial frequency of the nth clutter patch, expressed as
{ f c ; d , n = 2 v T r λ cos θ n cos φ n f c ; s , n = d λ cos ( θ n + θ e ) cos φ n
where φ n and θ n are the elevation and azimuth angles of the nth clutter patch, respectively, as shown in Figure 1, and λ denotes the signal wavelength.
According to Equation (1), assuming clutter and noise are independent of each other, the clutter-plus-noise covariance matrix (CNCM) can be obtained as
R I = R c + R n = E [ x c x c H ] + E [ x n x n H ] = n = 1 N c σ c ; n 2 [ v ( f c ; d , n , f c ; s , n ) v H ( f c ; d , n , f c ; s , n ) ] [ α ( n ) α H ( n ) ] + σ n 2 I N M
where E [ ] denotes expectation, [ ] H denotes conjugate transpose, and I N M denotes the unit matrix with a size of N M × N M .
To suppress clutter and detect moving targets, the output of STAP is the inner product of a space–time weighting vector w and the radar-received signal x , expressed as
y = w H x
To maintain the target power while minimizing the power of clutter and noise, the optimal weighting vector of the space–time filter can be calculated by
w opt = R I 1 v t / [ ( v t ) H R I 1 v t ] N M × 1
where ( ) 1 denotes matrix inverse and v t is the space–time steering vector of the target.
In practice, the CNCM of the RUT is unknown. In general, a certain number of training range cells that do not include the target are needed to estimate it. To do so, a typical method is to select some range cells near the RUT for training, whereas several range cells adjacent to the RUT on both sides are set as the guard cells to reduce the influence of target contamination [1]. Assuming that the training range cells are IID with the RUT, the CNCM of the RUT can be estimated via the sample matrix inversion (SMI) method [2], expressed as
R ^ I = 1 L l = 1 L x l x l H
where L is the number of IID training range cells and x l denotes the radar-received signal from the lth training range cell.
According to the RMB criterion [3], the output SCNR loss in dB of the SMI method compared to the optimal STAP method (i.e., the CNCM is known) can be expressed as
SCNR loss = 10 × log ( L O + 2 L + 1 )
where O = M N denotes the system DOF.
Equation (8) demonstrates that if the output SCNR loss is required to be less than 3 dB, the number of IID training range cells required by the SMI method is at least twice the system DOF, i.e., L 2 O , which is difficult to be satisfied in a practical non-uniform and non-stationary clutter environment.

2.2. SR-STAP

It can be seen from Equation (1) that, without considering the temporal and spatial non-ideal factors, the clutter signal can be viewed as the superposition of the space–time signals with different spatial and Doppler frequencies. Thus, by discretizing the spatial frequency domain and the Doppler frequency domain into N s = ρ s M and N d = ρ d N grids with N s N d N M , the clutter signal can be approximately expressed as
x c = i = 1 N d j = 1 N s γ i , j v ( f d , i , f s , j ) = Φ γ
where f d , i is the ith ( i = 1 , 2 , , N d ) Doppler frequency, f s , j is the jth ( j = 1 , 2 , , N s ) spatial frequency, v ( f d , i , f s , j ) is the space–time steering vector corresponding to the i-jth space–time grid, γ i , j denotes the complex amplitude of the i-jth space–time grid, γ = [ γ 1 , 1 , γ 2 , 1 , , γ N d , N s ] N s N d × 1 denotes the complex amplitude vector corresponding to all space–time grids, i.e., the space–time amplitude spectrum of the clutter, and Φ is a dictionary of space–time steering vectors, expressed as
Φ = [ v ( f d , 1 , f s , 1 ) , v ( f d , 2 , f s , 1 ) , , v ( f d , N d , f s , N s ) ] N M × N s N d
Based on Equation (9), the received signal of the lth training range cell without a target can be expressed as
x l = x c l + x n l = Φ γ l + x n l
Because of the space–time coupling property of clutter, its space–time amplitude spectrum is usually sparse. Hence, the SR-STAP method can estimate the space–time amplitude spectrum of clutter by solving a constrained optimization problem, expressed as
γ ^ l = arg min γ l γ l 0 , s . t . x l Φ γ l 2 ε
where | | | | 0 and | | | | 2 denote the L 0 norm and L 2 norm of a vector, respectively, and ε denotes the noise level.
With L training range cells, Equation (12) can be extended to the MMV model [15], expressed as
Γ ^ = arg min Γ Γ 2 , 0 , s . t . X Φ Γ F ε
where X = [ x 1 , x 2 , , x L ] N M × L , Γ = [ γ 1 , γ 2 , , γ L ] N s N d × L , | | | | 2 , 0 denotes the L 0 norm of the column vector obtained by the L 2 norm of each row of a matrix, and | | | | F denotes the Frobenius norm of a matrix.
By solving Equation (12) or (13) with a specific SR algorithm, such as FOCUSS, ADMM, or the FCSBL algorithm, the estimation of CNCM can be obtained as
R ^ I = 1 L l = 1 L i = 1 N d j = 1 N s | γ i , j , l | 2 v ( f d , i , f s , j ) v H ( f d , i , f s , j ) + σ n 2 I N M
where γ l , i , j denotes the complex amplitude of the i-jth space–time grid for the lth training range cell, i.e., the i-j-lth element of Γ .
Defining Z i , j = T ( Γ i , j ) = 1 L l = 1 L | γ i , j , l | 2 , Equation (14) can be rewritten as
R ^ I = i = 1 N d j = 1 N s Z i , j 2 v ( f d , i , f s , j ) v H ( f d , i , f s , j ) + σ n 2 I N M
Based on the estimated CNCM, the weighting vector of the space–time filter can be obtained by
w ^ opt = R ^ I 1 v t / [ ( v t ) H R ^ I 1 v t ] N M × 1
The SR-STAP method can obtain the estimation of CNCM using far fewer IID training range cells than the system DOF, i.e., L O . Hence, it has significant advantages over the SMI method in a practical environment.

3. DU-CG-STAP

Given the clutter spectrum SR estimation model in Equations (12) or (13), the performance of SR-STAP methods mainly depends on the adopted SR algorithm. Although many effective SR algorithms have been proposed, they have some common problems, e.g., parameter-setting difficulty and high computational complexity. In addition, in practical applications, various non-ideal factors will deteriorate the clutter sparsity and make the SR estimation model inaccurate, resulting in significant interferences deviating from the clutter ridge in the space–time domain and degrading the clutter suppression performance of SR-STAP methods. To solve these problems, a new STAP method, i.e., DU-CG-STAP, is proposed.
The main idea of DU-CG-STAP is to combine an SR-based DU-Net with an image-enhancement DNN. The SR-based DU-Net, namely 2DMA-Net, is used to obtain the clutter spectrum quickly from the raw radar data without parameter tuning. The image-enhancement DNN, namely CycleGAN, is used to process the clutter spectrum obtained by 2DMA-Net to generate an accurate and high-resolution counterpart.
The processing framework of the DU-CG-STAP method is shown in Figure 2. It realizes the nonlinear transform from the raw radar data X ˜ to the clutter spectrum Z ^ , i.e., Z ^ = ( X ˜ ) . The key to this method is the DU-CG network, where (1) the 2DMA-Net module is a solving network for the problem in Equation (13) with the network parameter as Θ A and the output as the clutter spectrum estimation Γ ˜ ^ N d × N s × L ; (2) the transform module T ( ) completes the single-channel processing of the spectrum Γ ˜ ^ N d × N s × L in the range dimension to obtain Y ^ 1 × N d × N s ; (3) the normalization module N ( ) normalizes the clutter spectrum Y ^ to obtain Y ^ N = N ( Y ^ ) = Y ^ / max ( Y ^ ) 1 × N d × N s as the input of GYZ; (4) the generator GYZ of CycleGAN is the clutter spectrum enhancement network with the parameter as Θ C and the output as the normalized clutter spectrum estimation Z ^ N 1 × N d × N s ; (5) the restoration module ( ) obtains Z ^ = ( Z ^ N ) = Z ^ N × max ( Y ^ ) 1 × N d × N s , i.e., the final output of the DU-CG network.
To summarize, the procedure of the DU-CG-STAP method is as follows.
Step 1. Implement the offline training of the DU-CG network (including 2DMA-Net and CycleGAN).
Step 2. Input the raw radar data into the trained DU-CG network to obtain the clutter spectrum estimation.
Step 3. Calculate the CNCM and the space–time weighting vector and then conduct clutter suppression and moving target detection.
In the following, the network structure, dataset construction method, and network training method of DU-CG will be introduced in detail.

3.1. Network Structure

3.1.1. 2DMA-Net

Since the L 2 , 0 norm is a discontinuous function, the complexity of directly solving Equation (13) is quite high. Thus, Equation (13) is usually solved by transforming it into an L 2 , 1 convex optimization problem, expressed as
Γ ^ = arg min Γ Γ 2 , 1 , s . t . X Φ Γ F ε
Introducing an auxiliary variable R N M × L , Equation (17) can be transformed into
{ Γ ^ , R ^ } = arg min Γ , R { Γ 2 , 1 + 1 2 ρ R F 2 } s . t . Φ Γ + R = X
where ρ > 0 denotes the regularization factor.
The augmented Lagrange function of Equation (18) is given by
{ Γ ^ , R ^ , Λ ^ } = arg min Γ , R , Λ Γ 2 , 1 + 1 2 ρ R F 2 + Λ , Φ Γ + R X + β 2 Φ Γ + R X F 2
where < , > denotes the inner product, Λ N M × L denotes the Lagrange multiplier, and β > 0 denotes the quadratic penalty factor.
Given an initial value { Γ ( 0 ) , R ( 0 ) , Λ ( 0 ) } , the MMV-ADMM algorithm solves Equation (19) by solving the following three sub-problems alternately with K iterations.
{ R ( k ) = arg min R 1 2 ρ R F 2 + β 2 Φ Γ ( k 1 ) + R X + Λ ( k 1 ) β F 2 Γ ( k ) = arg min Γ Γ 2 , 1 + β 2 Φ Γ + R ( k ) X + Λ ( k 1 ) β F 2 Λ ( k ) = Λ ( k 1 ) + β ( Φ Γ ( k ) + R ( k ) X )
where R ( k ) , Γ ( k ) , and Λ ( k ) denote the estimation of R , Γ , and Λ in the kth iteration ( k = 1 , 2 , , K ), respectively.
The solutions of Equation (20) can be expressed as [12,30]
{ R ( k ) = ρ β 1 + ρ β ( X Φ Γ ( k 1 ) Λ ( k 1 ) β ) Γ ( k ) = U ( k ) ( Γ ( k 1 ) + τ ρ β Φ H R ( k ) ) Λ ( k ) = Λ ( k 1 ) + β ( Φ Γ ( k ) + R ( k ) X )
where U = u × 1 1 × L N d N s × L , u = [ u 1 , 1 , u 2 , 1 , , u N d , N s ] T N d N s × 1 , u i , j = β Γ i , j 2 β Γ i , j 2 + τ , Γ i , j = [ γ i , j , 1 , γ i , j , 2 , , γ i , j , L ] , and τ is the iteration step size.
It can be seen from Equation (21) that the MMV-ADMM algorithm needs multiple matrix multiplications in each iteration, causing a high computing complexity. To improve the computing speed, the space–time 2D-decoupling process is implemented.
Firstly, the signal matrix X N M × L , noise matrix N N M × L , and clutter spectrum matrix Γ N s N d × L are space–time 2D-decoupled and transformed to the three-dimensional (3D) tensor form as X ˜ N × M × L , N ˜ N × M × L , and Γ ˜ N d × N s × L , respectively. Then, corresponding to the space–time dictionary Φ N M × N s N d , the spatial dictionary Φ s M × N s × 1 and the temporal dictionary Φ d N × N d × 1 in the 3D tensor form are constructed. At last, the radar-received signal tensor is expressed as
X ˜ = Φ d , Γ ˜ , Φ s T + N ˜
where · denotes the batch multiplication of multiple 3D tensors. For batch multiplication, the matrix slice of each tensor is taken from the third dimension for matrix multiplication. When the third-dimension size of a tensor is one, the batch multiplication takes the same matrix slice each time. For example, the batch multiplication of tensors a m × n × l , b n × p × l , and c p × q × 1 can be simply expressed as d = a , b , c m × q × l .
Based on Equation (22) and the batch multiplication process, the 2D-MMV-ADMM algorithm can be obtained from Equation (21), expressed as
{ R ˜ ( k ) = ρ β 1 + ρ β ( X ˜ Φ d , Γ ˜ ( k 1 ) , Φ s T Λ ˜ ( k 1 ) β ) N × M × L Γ ˜ ( k ) = U ˜ ( k ) ( Γ ˜ ( k 1 ) + τ ρ β Φ d H , R ˜ ( k ) , Φ s ) l N d × N s × L Λ ˜ ( k ) = Λ ˜ ( k 1 ) + β ( Φ d , Γ ˜ ( k ) , Φ s T + R ˜ ( k ) X ˜ ) N × M × L
where tensors R ˜ , U ˜ , and Λ ˜ are the space–time 2D-decoupled forms of R , U , and Λ , respectively.
Given the regularization factor ρ , the quadratic penalty factor β , and the iteration step τ in advance, the 2D-MMV-ADMM algorithm can obtain the clutter spectrum estimation as Γ ˜ ^ = Γ ˜ ( K ) . Then, the CNCM and the space–time weighting vector can be calculated according to Equations (15) and (16). It can be seen from Equations (21) and (23) that by using the number of multiplications in a single iteration as the indicator, the complexities of the MMV-ADMM algorithm and its space–time 2D-decoupled version are, respectively, O ( 2 N M N d N s L + ( N d N s ) 2 L + 3 N M L + N d N s L ) and O ( 2 N d N s L + N M ( N d + N s ) L + ( N + M ) N d N s L ) . Compared to the MMV-ADMM algorithm, the 2D-MMV-ADMM algorithm can significantly reduce the computational complexity.
However, in practical applications, the parameter setting for 2D-MMV-ADMM is usually difficult. Unreasonable parameter settings will affect the convergence performance, resulting in high computational complexity and low clutter spectrum estimation accuracy. To solve this problem, based on the idea of DU, the 2D-MMV-ADMM algorithm with K iterations is unfolded into a K-layer neural network, i.e., 2DMA-Net, as shown in Figure 3. The data-learning approach is used to obtain the optimal parameters for 2D-MMV-ADMM.
The input, output, and parameters of 2DMA-Net are the signal tensor X ˜ N × M × L , the clutter spectrum estimation Γ ˜ ^ = Γ ˜ ( K ) , and Θ A = { Θ A ( k ) } k = 1 K = { ρ k , β k , τ k } k = 1 K , respectively. The output of the k-th layer of 2DMA-Net is the Lagrange multiplier Λ ˜ ( k ) N × M × L , the auxiliary variable R ˜ ( k ) N × M × L , and the clutter spectrum Γ ˜ ( k ) N d × N s × L . With operations similar to Equation (23), the nonlinear function F k { } can be expressed as
{ Γ ˜ ( k ) , R ˜ ( k ) , Λ ˜ ( k ) } = F k { X ˜ , Φ d , Φ s , Γ ˜ ( k 1 ) , R ˜ ( k 1 ) , Λ ˜ ( k 1 ) , Θ A ( k ) }
2DMA-Net is driven by both data training and the theoretical model, hence having the advantages of data adaptability and model interpretability. With optimized network parameters, 2DMA-Net can achieve a higher convergence performance than the 2D-MMV-ADMM algorithm, thus reducing the computing complexity and improving the performance for estimating the clutter spectrum.

3.1.2. CycleGAN

In practice, non-ideal factors will reduce the clutter spectrum estimation accuracy of 2DMA-Net, resulting in some interferences in the space–time domain. To solve this problem, CycleGAN is used as an image-enhancement mapping tool to process the low-accuracy clutter spectrum output of 2DMA-Net. The processing framework of CycleGAN is shown in Figure 4, where the unpaired low-accuracy clutter spectrum Y p and high-accuracy clutter spectrum Z p are both the input data. There are two generators of CycleGAN, GYZ and GZY, where GYZ maps the low-accuracy clutter spectrum Y p into the high-accuracy domain to obtain Z ^ p and GZY maps the high-accuracy clutter spectrum Z p into the low-accuracy domain to obtain Y ^ p . Discriminators DY and DZ improve the mapping capability of the generators continuously in an adversarial mechanism. After training, the generator GYZ of CycleGAN has the high-accuracy mapping capability for the low-accuracy clutter spectrum.
The network structures of the generator and discriminator of the CycleGAN used in this study are shown in Figure 5, where sigmoid(·) denotes the sigmoid function, Tanh(·) denotes the hyperbolic tangent function, Conv2d denotes the 2D convolution process with the convolution kernel dimension as c e × f e × f e × n e , c e denotes the number of input channels, f e denotes the length and width of the convolution kernel, n e denotes the number of convolution kernels (i.e., the number of output channels), Residual Block denotes the cascade of two Conv2d layers, and unlike Conv2d that implements the image down-sampling process, ConvTranspose2d implements the image up-sampling process to expand the image size. It should be noted that to better conduct clutter spectrum enhancement tasks and maintain low computational complexity, some appropriate modifications are made to the original network structures of CycleGAN given in [29].

3.2. Dataset Construction

Compared to supervised learning, without using paired training data, unsupervised learning and self-supervised learning can acquire a large number of training data at a low cost. In this study, 2DMA-Net uses a self-supervised learning method and CycleGAN uses an unsupervised learning method, for which the training dataset is constructed with the following three steps.
  • Step 1. Parameter setting
First, some parameters of the airborne radar system, i.e., signal wavelength λ , pulse repetition interval T r , ULA element number M , element spacing d , CPI pulse number N , and the training range cell number L , are fixed. In addition, it is set that each range cell consists of N c clutter patches that are uniformly distributed in the azimuth angle range [ 0 , π ] . The noise power is fixed to σ n 2 = 1 and the scattering coefficients of clutter patches obey a complex Gaussian distribution with the amplitude determined by the clutter-to-noise ratio (CNR).
Then, some intermediate parameters are calculated. With c as the speed of light, the maximum unambiguous range is calculated as R u = c T r / 2 . Given the space–time frequency range [ f s , min , f s , max ] and [ f d , min , f d , max ] and the grid number N d = ρ d N and N s = ρ s M , the spatial dictionary Φ s M × N s × 1 and the temporal dictionary Φ d N × N d × 1 are, respectively, constructed.
Finally, to mimic complicated scenarios, other parameters used to obtain the raw radar data are assumed to be uniformly randomly distributed within specified ranges, i.e., the airplane height H U [ H min , H max ] , the airplane velocity v U [ v min , v max ] , the non-side-looking angle θ e U [ θ e , min , θ e , max ] , the detection range R 0 U [ R min , R u ] , the ICM σ v U [ σ v , min , σ v , max ] , the array element amplitude error σ a U [ σ a , min , σ a , max ] , the array element phase error σ p U [ σ p , min , σ p , max ] , and the CNR CNR U ( CNR min , CNR max ) .
  • Step 2. Data generating
According to the above settings, P different scenarios with random parameters are simulated and the raw radar data X ˜ p ( p = 1 , 2 , , P ) corresponding to each scenario are generated based on Equation (1) and used as the input data for 2DMA-Net. After training 2DMA-Net with the self-supervised method as presented in the following subsection, the low-accuracy clutter spectrum Y p is generated corresponding to each set of raw radar data X ˜ p and used as the input data for CycleGAN.
To train the CycleGAN with the unsupervised method, P different scenarios are simulated with random parameters. Meanwhile, the theoretical CNCM is calculated for each scenario based on Equation (4), where no array amplitude/phase error is contained and thus the spatial weighting vector is fixed as α s ( n ) = I M × 1 . Then, based on the minimum variance distortionless response (MVDR) algorithm [31], the high-accuracy clutter spectrum Z p corresponding to each scenario is generated and is also used as the input data for CycleGAN.
  • Step 3. Dataset partitioning
In Step 2, the generated dataset for 2DMA-Net is { X ˜ p } p = 1 P and the generated dataset for CycleGAN is { Y p , Z p } p = 1 P . As shown in Figure 6, in this step, the generated datasets are divided into training datasets and validation datasets according to a certain proportion, with the sizes as P train and P test , respectively.

3.3. Network Training

3.3.1. 2DMA-Net

In most existing DU-Nets, the supervised training method is used, i.e., the output label for each input data is prepared for network training. However, for airborne radar STAP applications, the output label of 2DMA-Net is difficult to obtain as no exact clutter spectrum is available for each set of input raw radar data. A possible solution is to apply the 2D-MMV-ADMM algorithm with fixed manual-tuned parameters and sufficient iterations to solve Equation (22) to obtain the clutter spectrum estimation as the output label for 2DMA-Net. However, with fixed parameters this method cannot guarantee the estimation accuracy for different inputs, resulting in the distortion of output labels. In addition, to obtain convergence, this method needs a lot of iterations, resulting in high computing costs. To solve these problems, the self-supervised training method is adopted by 2DMA-Net without preparing output labels.
With the output clutter spectrum estimation Γ ˜ ^ p of 2DMA-Net for each set of input raw radar data X ˜ p , the clutter data can be reconstructed as X ˜ ^ p = Φ d , Γ ˜ ^ p , Φ s T . Then, the following network loss function is defined for the self-supervised training of 2DMA-Net.
( Θ A ) = 1 P train p = 1 P train ( R ^ I , p R I , p F 2 + α Γ ˜ ^ p 1 )
where R I , p = X p X p H / L , R ^ I , p = X ^ p X ^ p H / L , X p and X ^ p are the matrix forms of the tensors X ˜ p and X ˜ ^ p , and α is a constant.
It should be noted that in Equation (25), two functions are combined to define the network loss function of 2DMA-Net. The first function (i.e., MSE loss function) is used to ensure the estimation accuracy of the clutter spectrum with the consideration that the more accurate the estimation of Γ ˜ ^ p , the smaller the difference between X ˜ p ( R I , p ) and X ˜ ^ p ( R ^ I , p ). The second function (i.e., L1 regularization loss function) is used to improve the sparsity of the clutter spectrum estimation. If only the MSE loss function is used, the clutter spectrum estimation results may be quite different from the sparse solution of the SR-STAP problem. If only the L1 regularization loss function is used, as the clutter sparsity and the SR estimation model will be seriously damaged by the non-ideal factors (e.g., ICM, element amplitude/phase error, and low CNR), the performance of 2DMA-Net may degrade a lot with significant interferences in the space–time domain. Hence, by using a balancing coefficient α, the MSE and L1 regularization combined loss function is used by 2DMV-Net for network training to achieve a high clutter spectrum estimation performance.
Based on the loss function given in Equation (25), with the parameters of each 2DMA-Net layer initialized as Θ A = { ρ k = ρ 0 , β k = β 0 , τ k = β 0 } k = 1 K , the optimal parameters of 2DMA-Net Θ A * = { ρ k * , β k * , τ k * } k = 1 K can be obtained via the back-propagation method [32,33], expressed as
Θ A * = arg   min Θ A ( Θ A )

3.3.2. CycleGAN

For DNNs using the supervised training method, a paired dataset is required. For airborne radar STAP applications, the practical clutter environment is usually unknown in advance, resulting in difficulties for proper dataset construction. To solve this problem, an unsupervised training method is adopted by CycleGAN. To realize the mutual mapping between the clutter spectra in the low-accuracy domain and the high-accuracy domain via the unpaired dataset, CycleGAN conducts the unsupervised training based on an adversarial criterion and cycle-consistency criterion, which are detailed as follows.
(1)
Adversarial training
Consider that the generator GYZ can accurately map the low-accuracy clutter spectrum Y p to the high-accuracy domain to obtain Z ^ p = G Y Z ( Y p ) (namely the fake high-accuracy clutter spectrum). Then, it will be difficult for the discriminator DZ to distinguish Z ^ p from the true high-accuracy clutter spectrum dataset { Z p } p = 1 P . The adversarial training process will continuously improve the discriminating capability of DZ on the fake and true spectrum and based on the feedback of DZ, GYZ will continuously improve its high-accuracy mapping capability on the low-accuracy clutter spectrum. Thus, the following loss function is defined for the generator GYZ and the discriminator DZ, expressed as
GAN ( G Y Z , D Z ) = E [ logD Z ( Z p ) ] + E [ log ( 1 D Z ( G Y Z ( Y p ) ) ) ]
where E [ ] denotes expectation and logD Z ( Z p ) and log ( 1 D Z ( G Y Z ( Y p ) ) ) denote the probabilities that the true and fake high-accuracy clutter spectra can be correctly discriminated by DZ, respectively.
Similarly, the following loss function is defined for the generator GZY and the discriminator DY, expressed as
GAN ( G Z Y , D Y ) = E [ logD Y ( Y p ) ] + E [ log ( 1 D Y ( G Z Y ( Z p ) ) ) ]
where logD Y ( Y p ) and log ( 1 D Y ( G Z Y ( Z p ) ) ) denote the probability that the true and fake low-accuracy clutter spectra can be correctly discriminated by DY, respectively.
The training process based on the adversarial criterion optimizes the generators GYZ/GZY and the discriminators DY/DZ simultaneously, expressed as min G Y Z max D Z GAN ( G Y Z , D Z ) and min G Z Y max D Y GAN ( G Z Y , D Y ) , i.e., the generators and discriminators will oppositely minimize and maximize the same loss function.
(2)
Cycle-consistency training
The goal of adversarial training is to make it difficult for DZ to discriminate Z ^ p from the true high-accuracy clutter spectrum dataset { Z p } p = 1 P . However, it cannot guarantee that Y p and Z ^ p = G Y Z ( Y p ) correspond to the same situation. For example, the low-accuracy clutter spectrum in the side-looking case may be transformed by GYZ into a high-accuracy clutter spectrum in the non-side-looking case. In other words, the adversarial training process only forces Z ^ p to belong to the high-accuracy domain but cannot ensure that Z ^ p is the real desired high-accuracy clutter spectrum counterpart of Y p .
Based on the cycle-consistency criterion, if Y p can be recovered to the original data by GYZ and GZY successively, i.e., G Z Y ( G Y Z ( Y p ) ) Y p , it can guarantee Z ^ p and Y p correspond to the same situation. Similarly, for Z p , it has G Y Z ( G Z Y ( Z p ) ) Z p . Hence, the following loss function is defined for the cycle-consistency training, expressed as
cyc   ( G Y Z , G Z Y ) = E [ G Z Y ( G Y Z ( Y p ) ) Y p 1 ] + E [ G Y Z ( G Z Y ( Z p ) ) Z p 1 ]
(3)
Full training
To ensure the mapping and correspondence of the clutter spectrum at the same time, the full training process is conducted. Combining the adversarial loss function and the cycle-consistency loss function with their importance balanced by a coefficient μ , the full loss function for CycleGAN is defined as
( G Y Z , G Z Y , D Z , D Y , Θ C ) = GAN ( G Z Y , D Y ) + GAN ( G Y Z , D Z ) + μ cyc   ( G Y Z , G Z Y )
Then, by using the Glorot method [34,35] for initialization, the optimal network parameters of CycleGAN can be obtained via the back-propagation method, expressed as
Θ C * = arg min G Y Z , G Z Y max D Z , D Y ( G Y Z , G Z Y , D Z , D Y , Θ C )

4. Experiment Results

In this section, the performance of the proposed DU-CG-STAP method is verified and compared with three typical SR-STAP methods, i.e., MMV-FOCUSS-STAP, MMV-FCSBL-STAP, and MMV-ADMM-STAP, via various simulations with the parameters shown in Table 1, which are set according to their typical values [13,15,21].
In MMV-FOCUSS-STAP, the number of iterations is set as 200 and the sparsity parameter is set as 0.2. In MMV-FCSBL-STAP, the number of iterations is set as 30 and the noise variance is initialized as 10−5. In MMV-FCSBL-STAP, the parameters are set as ρ = 0.5 , β = 0.2 , τ = 0.04 , and K = 2000 . In the self-supervised training of 2DMA-Net, the coefficient of the L1 regularization loss function, the number of network layers, the initial learning rate, and the training epoch are set as α = 0.01 , K = 30 , 10−4, and 500, and the parameters of each layer are initialized as Θ A = { ρ k = 0.5 , β k = 0.2 , τ k = 0.04 } k = 1 30 . In the unsupervised training of CycleGAN, the coefficient of the cycle-consistency loss, the initial learning rate, and the training epoch are set as μ = 10 , 2 × 10−5, and 500, respectively.

4.1. Network Convergence Analysis

In this subsection, the convergence results of network training are presented. Figure 7a shows the combined loss of 2DMA-Net, the cycle-consistency loss, and the full loss of CycleGAN during the training process. It can be seen that the losses decrease gradually and remain unchanged from about 200 epochs, demonstrating the favorable convergence performance of 2DMA-Net and CycleGAN. Figure 7b shows the discrimination probability curves of the discriminator DZ on the true spectrum and the fake spectrum. The discrimination probability of 1 indicates that the discrimination results are true and the discrimination probability of 0 indicates that the discrimination results are fake. It can be seen that the discrimination probability of DZ simultaneously increases to 1 on the true spectrum and decreases to 0 on the fake spectrum. The increasing capacity of the discriminator DZ to distinguish between the true and fake spectrums indicates the increasing capacity of the generator GYZ to map the low-accuracy clutter spectrum to the high-accuracy clutter spectrum, hence increasing the following CNCM estimation accuracy.

4.2. Clutter Spectrum Estimation

In this subsection, the clutter spectrum estimation results of the proposed DU-CG network are presented under different situations. For comparison, the results obtained via MMV-FOCUSS, MMV-FCSBL, and MMV-ADMM are also shown. As a reference, the MVDR clutter spectrum is calculated based on the theoretical CNCM.
First, Figure 8 shows the estimation results using different methods in the ideal case, i.e., the case with the clutter ridge slope as 1, non-side-looking angle as 0, and no ICM or element amplitude/phase error. It can be seen that as the clutter has a high sparsity in the ideal case, these methods can all estimate the clutter spectrum accurately. As a module of DU-CG, the results obtained using 2DMA-Net have relatively low accuracy where the clutter ridge is broadened. However, as the clutter feature is clearly achieved, based on the output of 2DMA-Net, the CycleGAN in DU-CG can successfully obtain a high-accuracy clutter spectrum estimation.
Then, Figure 9 and Figure 10 show the estimation results in the non-ideal cases, where the clutter ridge slope is changed to 1.34 and the non-side-looking angle is changed to 16.5°, respectively. It can be seen that, as the clutter sparsity deteriorates in these two cases, the estimation accuracy of typical MMV-SR algorithms degrades significantly. The clutter ridges obtained by these algorithms are broadened and some significant interferences deviating from the clutter ridges are generated. 2DMA-Net can obtain the low-accuracy clutter spectrum estimation with clear clutter features, and thus, based on the output of 2DMA-Net, a high-accuracy clutter spectrum estimation can be obtained by CycleGAN, which is consistent with the reference. These results demonstrate that in the non-ideal cases, typical MMV-SR algorithms are seriously affected by the deteriorated clutter sparsity, whereas the proposed model-driven and data-driven DU-CG network can effectively overcome this problem and adaptively extract the clutter feature to obtain the high-accuracy clutter spectrum estimation.
Furthermore, Figure 11 and Figure 12 show the estimation results of DU-CG in the other two non-ideal cases, where the ICM is set as 0.5 m/s and the element amplitude/phase error is set as 0.14/3.83°. It can be seen that in the presence of ICM, the clutter is broadened along the Doppler dimension due to the temporal decorrelation problem, leading to the damaged clutter sparsity. Hence, typical MMV-SR algorithms will have decreased clutter spectrum estimation accuracy. In the presence of an array element amplitude/phase error, as the SR estimation model and the clutter sparsity are both damaged, the performance of typical MMV-SR algorithms degrades significantly. However, although the performance of 2DMA-Net also degrades in these two cases, the clutter feature is maintained. Then, as CycleGAN can reduce the width of the clutter ridge and suppress the discrete interferences in the space–time domain, a high-accuracy clutter spectrum closest to the reference can still be obtained by the DU-CG network.
Finally, all the above-mentioned non-ideal factors are considered, giving the results shown in Figure 13, where the clutter ridge slope is 0.67, the non-side-looking angle is 15.50°, the ICM is 0.24 m/s, and the array element amplitude/phase error is 0.10/4°. The results show that in such a complicated case, the performance of typical MMV-SR algorithms degrades significantly, the clutter ridge feature distorts severely, and a lot of false peaks appear in the space–time domain. As the proposed DU-CG network can adaptively acquire the clutter features and filter out the interferences caused by non-ideal factors, an accurate estimation of the clutter spectrum is still achieved.

4.3. Clutter Suppression Performance

In this subsection, the clutter suppression performance of different STAP methods is compared using the SCNR loss as the indicator. Keeping the spatial frequency of the target as 0 and linearly varying its normalized Doppler frequency in the range [ 0.5 , 0.5 ] , the obtained results are shown in Figure 14, where the subfigures (a)–(f), respectively, correspond to Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.
The comparison in Figure 14a shows that, in the ideal case, the MMV-FOCUSS-STAP method and MMV-FCSBL-STAP method can achieve the best clutter suppression performance, whereas the proposed DU-CG-STAP method can obtain slightly worse suboptimal performance, which is better than the MMV-ADMM-STAP method. The comparisons in Figure 14b,c show that the clutter suppression performance of typical MMV-SR-STAP methods degrades with the clutter sparsity deterioration, which is manifested by the broadened notch in the zero-Doppler region and the false notches deviating from the clutter ridge. The proposed DU-CG-STAP method can obtain a narrower clutter suppression notch and avoid false notches. The comparisons in Figure 14d,e show that in the presence of ICM and array element amplitude/phase error, typical MMV-SR-STAP methods have significant SCNR losses in almost the entire Doppler frequency range, hence they will suppress not only the clutter but also the target, resulting in low target-detection performance. The proposed DU-CG-STAP method can form an effective suppression notch for the clutter and maintain the power for the target, hence it has a higher performance. The comparison in Figure 14f shows that under conditions with all considered non-ideal factors, compared to typical MMV-SR-STAP methods, the proposed DU-CG-STAP method can still obtain a high clutter suppression performance that is close to the theoretical optimal STAP.

4.4. Computational Complexity Analysis

In this subsection, the computational complexity of the DU-CG network is analyzed and compared with MMV-FOCUSS, MMV-FCSBL, and MMV-ADMM. It should be noted that when applied, the computational complexity of DU-CG is the sum of those of 2DMA-Net and the generator GYZ in CycleGAN. With the only difference in the iterative parameters, the computations of 2DMA-Net and the 2D-MMV-ADMM algorithm are the same. Thus, with the same number of network layers and iterations, 2DMA-Net and the 2D-MMV-ADMM algorithm will have the same computational complexity. Using the multiplication numbers as the indictor, the computational complexities of different algorithms are given in Table 2.
According to Table 2, the computational complexities of different algorithms under different conditions are shown in Figure 15. Figure 15a corresponds to the conditions of N = N d / 5 = 10 , M = N s / 5 = 10 , and the number of algorithm iterations or network layers K varying from 20 to 50 with a step of 5. Figure 15b corresponds to the conditions of M = N = N d / 5 = N s / 5 varying from 5 to 30 with a step of 5 and the number of algorithm iterations or network layers K of MMV-FOCUSS, MMV-FCSBL, MMV-ADMM, and the DU-CG network as 200, 30, 2000, and 30 (which are determined considering their convergence performance). The comparisons show that the computational complexities of 2DMA-Net and the generator GYX in CycleGAN are much lower than the other methods. Hence, the proposed DU-CG network can always obtain a faster convergence speed under different conditions.

4.5. Rationality of DU-CG-STAP

In the CycleGAN of DU-CG-STAP, the discriminator DZ learns the features of the high-accuracy clutter spectrum with the capacity to discriminate the true and fake spectra, which is continuously improved based on the unpaired dataset. At the same time, the generator GYZ is committed to mapping the low-accuracy clutter spectrum into the high-accuracy domain. If the low-accuracy clutter spectrum is provided with poor quality, it will be difficult for CycleGAN to extract the clutter features and complete the high-accuracy reconstruction task in the unsupervised training process. Hence, to illustrate the rationality of the DU-CG-STAP processing framework, the following results are provided.
In the proposed processing framework, the low-accuracy clutter spectrum dataset is generated by the self-supervised trained 2DMA-Net and used as the input data for CycleGAN. If the low-accuracy clutter spectrum dataset is generated by some low-accuracy and low-resolution methods, the high-resolution clutter spectrum reconstruction performance of CycleGAN will seriously degrade. For example, with the low-accuracy clutter spectrum obtained using the Fourier transform and the MVDR methods that were conducted on the raw radar data, the low-accuracy and high-accuracy mapping results of CycleGAN under different conditions are obtained and shown in Figure 16a,b, where the same network scale and training process with the proposed method are used.
It can be seen in Figure 16a that due to the high sidelobes in the Fourier clutter spectrum, the generator GYZ of CycleGAN incorrectly extracts many high values, resulting in significant distortions of the clutter features. It can be seen in Figure 16b that with a small amount of training range cells, the clutter ridge obtained by the MVDR algorithm broadens and some noises exist in the clutter spectrum. Hence, even though the generator GYZ can extract the clutter features, it cannot effectively reduce the clutter ridge width and suppress the noisy spectrum component.
On the contrary, by generating the low-accuracy clutter spectrum dataset via the MMV-ADMM algorithm and the trained 2DMA-Net, the low-accuracy and high-accuracy mapping results of CycleGAN are obtained and shown in Figure 16c,d. Since the clutter spectra obtained by these two approaches have no obvious sidelobe/noises and the features of the clutter ridge are clear, CycleGAN can filter out the interferences caused by non-ideal factors in the unsupervised training process so as to complete the high-accuracy clutter spectrum reconstruction task.

5. Conclusions

To solve the problems of high computational complexity, parameter-setting difficulties, and degraded performance caused by non-ideal factors in the conventional SR-STAP methods for airborne radar moving-target detection, a novel DU-CG-STAP method has been proposed in this paper. The processing framework, network structure, dataset construction, and training methods of the proposed method have been introduced in detail. The simulation results obtained under different situations have shown that compared to existing typical SR-STAP methods, the proposed method can simultaneously improve the clutter spectrum estimation accuracy and reduce the computational complexity, thus achieving a higher clutter suppression and target detection performance. In future work, we will focus on the improvement of the SR-based DU-Nets and the image-enhancement DNNs for STAP applications.

Author Contributions

Conceptualization, B.Z., W.F. and X.W.; methodology, B.Z., X.W. and W.F.; software, B.Z., H.Z. and F.L.; validation, B.Z., H.Z. and F.L.; formal analysis, W.F. and X.W.; investigation, W.F. and X.W.; resources, W.F. and H.Z.; data curation, F.L. and B.Z.; writing—original draft preparation, B.Z.; writing—review and editing, W.F. and X.W.; visualization, B.Z., H.Z. and F.L.; supervision, W.F. and H.Z.; project administration, W.F.; funding acquisition, W.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China, No. 62001507, and the Young Talent fund of University Association for Science and Technology in Shaanxi, China, No. 20210106.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guerci, J.R. Space-Time Adaptive Processing for Radar; Artech House: London, UK, 2014. [Google Scholar]
  2. Brennan, L.E.; Reed, I.S. Theory of adaptive radar. IEEE Trans. Aerosp. Electron. Syst. 1973, AES-9, 237–252. [Google Scholar] [CrossRef]
  3. Reed, I.S.; Mallett, J.D.; Brennan, L.E. Rapid convergence rate in adaptive arrays. IEEE Trans. Aerosp. Electron. Syst. 1974, AES-10, 853–863. [Google Scholar] [CrossRef]
  4. Melvin, W.L. A STAP overview. IEEE Aerosp. Electron. Syst. Mag. 2004, 19, 19–35. [Google Scholar] [CrossRef]
  5. Corbell, P.M.; Perez, J.J.; Rangaswamy, M. Enhancing GMTI performance in non-stationary clutter using 3D STAP. In Proceedings of the IEEE Radar Conference, Waltham, MA, USA, 17–20 April 2007; IEEE: New York, NY, USA, 2007; pp. 647–652. [Google Scholar]
  6. Wang, Y.L.; Chen, J.W.; Bao, Z.; Peng, Y.N. Robust space-time adaptive processing for airborne radar in nonhomogeneous clutter environments. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 70–81. [Google Scholar] [CrossRef]
  7. Peckham, C.D.; Haimovich, A.M.; Ayoub, T.F.; Goldstein, J.; Reid, I.S. Reduced-rank STAP performance analysis. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 664–676. [Google Scholar] [CrossRef]
  8. Klemm, R. Adaptive clutter suppression for airborne phased array radars. In IEE Proceedings F (Communications, Radar and Signal Processing); IET Digital Library: Stevenage, UK, 1983; Volume 130, pp. 125–132. [Google Scholar]
  9. Sarkar, T.K.; Wang, H.; Park, S.; Adve, R.; Koh, J.; Kim, K.; Brown, R.D. A deterministic least squares approach to space time adaptive processing (STAP). Trans. Antennas Propag. 2001, 49, 91–103. [Google Scholar] [CrossRef] [Green Version]
  10. Yang, Z.; Li, X.; Wang, H.; Jiang, W. On clutter sparsity analysis in space–time adaptive processing airborne radar. IEEE Geosci. Remote Sens. Lett. 2013, 10, 1214–1218. [Google Scholar] [CrossRef]
  11. Sun, K.; Meng, H.; Wang, Y.; Wang, X. Direct data domain STAP using sparse representation of clutter spectrum. Signal Process. 2011, 91, 2222–2236. [Google Scholar] [CrossRef] [Green Version]
  12. Yang, Z.C.; Lamare, R.; Liu, W. Sparsity-based STAP using alternating direction method with gain/phase errors. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 2756–2768. [Google Scholar] [CrossRef]
  13. Wang, Z.; Xie, W.; Duan, K.; Wang, Y. Clutter suppression algorithm based on fast converging sparse Bayesian learning for airborne radar. Signal Process. 2017, 130, 159–168. [Google Scholar] [CrossRef]
  14. Feng, W.; Guo, Y.; Zhang, Y.; Gong, Y. Airborne radar space time adaptive processing based on atomic norm minimization. Signal Process. 2018, 148, 31–40. [Google Scholar] [CrossRef]
  15. Duan, K.; Wang, Z.; Xie, W.; Chen, H.; Wang, Y. Sparsity-based STAP algorithm with multiple measurement vectors via sparse Bayesian learning strategy for airborne radar. IET Signal Process. 2017, 11, 544–553. [Google Scholar] [CrossRef]
  16. Li, Z.; Ye, H.; Liu, Z.; Sun, Z.; An, H.; Wu, J.; Yang, J. Bistatic SAR Clutter-Ridge Matched STAP Method for Non-stationary Clutter Suppression. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–14. [Google Scholar]
  17. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  18. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 12299–12310. [Google Scholar]
  19. Gao, J.; Deng, B.; Qin, Y.; Wang, H.; Li, X. Enhanced radar imaging using a complex-valued convolutional neural network. IEEE Geosci. Remote Sens. Lett. 2018, 16, 35–39. [Google Scholar] [CrossRef] [Green Version]
  20. Hu, X.; Feng, W.; Guo, Y.; Wang, Q. Feature learning for SAR target recognition with unknown classes by using CVAE-GAN. Remote Sens. 2021, 13, 3554. [Google Scholar] [CrossRef]
  21. Duan, K.; Chen, H.; Xie, W.; Wang, Y. Deep learning for high-resolution estimation of clutter angle-Doppler spectrum in STAP. IET Radar Sonar Navig. 2022, 16, 193–207. [Google Scholar] [CrossRef]
  22. Gregor, K.; Lecun, Y. Learning fast approximations of sparse coding. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 399–406. [Google Scholar]
  23. Borgerding, M.; Schniter, P.; Rangan, S. AMP-inspired deep networks for sparse linear inverse problems. IEEE Trans. Signal Process. 2017, 65, 4293–4308. [Google Scholar] [CrossRef]
  24. Liu, J.; Chen, X. ALISTA: Analytic weights are as good as learned weights in LISTA. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  25. Yang, Y.; Sun, J.; Li, H.; Xu, Z. ADMM-CSNet: A deep learning approach for image compressive sensing. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 521–538. [Google Scholar] [CrossRef]
  26. Zhu, H.; Feng, W.; Feng, C.; Zou, B.; Lu, F. Deep Unfolding Based Space-Time Adaptive Processing Method for Airborne Radar. J. Radars 2022, 11, 1–16. [Google Scholar]
  27. Hu, X.; Xu, F.; Guo, Y.; Feng, W.; Jin, Y.Q. MDLI-Net: Model-driven learning imaging network for high-resolution microwave imaging with large rotating angle and sparse sampling. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–17. [Google Scholar] [CrossRef]
  28. Zhao, S.; Ni, J.; Liang, J.; Xiong, S.; Luo, Y. End-to-End SAR Deep Learning Imaging Method Based on Sparse Optimization. Remote Sens. 2021, 13, 4429. [Google Scholar] [CrossRef]
  29. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  30. Sun, L.; Liu, J.; Chen, J.; Ye, J. Efficient recovery of jointly sparse vectors. Adv. Neural Inf. Process. Syst. 2009, 22, 1812–1820. [Google Scholar]
  31. Capon, J. High-resolution frequency-wavenumber spectrum analysis. Proc. IEEE 1969, 57, 1408–1418. [Google Scholar] [CrossRef] [Green Version]
  32. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  33. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  34. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics—JMLR Workshop and Conference Proceedings, Sardinia, Italy, 13–15 May 2010; Microtome Publishing: Brookline, MA, USA, 2010; pp. 249–256. [Google Scholar]
  35. Kumar, S.K. On weight initialization in deep neural networks. arXiv 2017, arXiv:1704.08863. [Google Scholar]
Figure 1. Geometry model of airborne radar.
Figure 1. Geometry model of airborne radar.
Remotesensing 14 03472 g001
Figure 2. Processing framework of DU-CG-STAP.
Figure 2. Processing framework of DU-CG-STAP.
Remotesensing 14 03472 g002
Figure 3. Network structure of 2DMA-Net.
Figure 3. Network structure of 2DMA-Net.
Remotesensing 14 03472 g003
Figure 4. Processing framework of CycleGAN.
Figure 4. Processing framework of CycleGAN.
Remotesensing 14 03472 g004
Figure 5. Network structures of the generator and discriminator of the CycleGAN in this study.
Figure 5. Network structures of the generator and discriminator of the CycleGAN in this study.
Remotesensing 14 03472 g005
Figure 6. Partitioning method of the generated datasets.
Figure 6. Partitioning method of the generated datasets.
Remotesensing 14 03472 g006
Figure 7. Network training convergence results: (a) training loss curves of 2DMA-Net and CycleGAN and (b) discrimination probability curves of the discriminator GZ.
Figure 7. Network training convergence results: (a) training loss curves of 2DMA-Net and CycleGAN and (b) discrimination probability curves of the discriminator GZ.
Remotesensing 14 03472 g007
Figure 8. Clutter spectrum estimation results obtained using different methods in the ideal case.
Figure 8. Clutter spectrum estimation results obtained using different methods in the ideal case.
Remotesensing 14 03472 g008aRemotesensing 14 03472 g008b
Figure 9. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 1.34.
Figure 9. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 1.34.
Remotesensing 14 03472 g009
Figure 10. Clutter spectrum estimation results obtained using different methods in the non-ideal case with a non-side-looking angle of 16.5°.
Figure 10. Clutter spectrum estimation results obtained using different methods in the non-ideal case with a non-side-looking angle of 16.5°.
Remotesensing 14 03472 g010
Figure 11. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the ICM as 0.5 m/s.
Figure 11. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the ICM as 0.5 m/s.
Remotesensing 14 03472 g011aRemotesensing 14 03472 g011b
Figure 12. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the element amplitude/phase error as 0.14/3.83°.
Figure 12. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the element amplitude/phase error as 0.14/3.83°.
Remotesensing 14 03472 g012
Figure 13. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 0.67, the non-side-looking angle as 15.50°, the ICM as 0.24 m/s, and the array element amplitude/phase error as 0.10/4°.
Figure 13. Clutter spectrum estimation results obtained using different methods in the non-ideal case with the clutter ridge slope as 0.67, the non-side-looking angle as 15.50°, the ICM as 0.24 m/s, and the array element amplitude/phase error as 0.10/4°.
Remotesensing 14 03472 g013
Figure 14. SCNR loss curves of different methods for a target with zero spatial frequency and varying Doppler frequency; subfigures (af) correspond to Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, respectively.
Figure 14. SCNR loss curves of different methods for a target with zero spatial frequency and varying Doppler frequency; subfigures (af) correspond to Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13, respectively.
Remotesensing 14 03472 g014
Figure 15. Computational complexities of different algorithms under different conditions: (a) the computational complexity varies with the iteration number and (b) the computational complexity varies with the array element number.
Figure 15. Computational complexities of different algorithms under different conditions: (a) the computational complexity varies with the iteration number and (b) the computational complexity varies with the array element number.
Remotesensing 14 03472 g015
Figure 16. Low-accuracy and high-accuracy mapping results of CycleGAN with the low-accuracy clutter spectrum dataset obtained by (a) Fourier transform, (b) MVDR algorithm, (c) MMV-ADMM algorithm, and (d) the trained 2DMA-Net.
Figure 16. Low-accuracy and high-accuracy mapping results of CycleGAN with the low-accuracy clutter spectrum dataset obtained by (a) Fourier transform, (b) MVDR algorithm, (c) MMV-ADMM algorithm, and (d) the trained 2DMA-Net.
Remotesensing 14 03472 g016
Table 1. Simulation parameters.
Table 1. Simulation parameters.
ParameterNotationValue
Element spacing d 0.1   m
Signal wavelength λ 0.2   m
Pulse repetition interval T r 0.5   ms
ULA Element number M 10
CPI Pulse number N 10
Training range cell number L 2
Spatial frequency range [ f s , min , f s , max ] [ 0.5 , 0.5 ]
Doppler frequency range [ f d , min , f d , max ] [ 0.5 , 0.5 ]
Number of spatial frequencies N s 50
Number of Doppler frequencies N d 50
Number of clutter patches N c 181
Airplane height H U [ 8 , 15 ]   km
Airplane velocity v U [ 70 , 120 ]   m / s
Detection range R 0 U [ 15   km , R u ]
Non-side-looking angle θ e U [ 30 , 30 ]  
Clutter-to-noise-ratio CNR U [ 30 , 50 ]   dB
Internal clutter motion σ v U [ 0 , 1 ]   m / s
Element amplitude error σ a U [ 0 , 0.2 ]
Element phase error σ p U [ 0 , 10 ]  
Size of training dataset P train 10 , 000
Size of validation dataset P test 2000
Table 2. Computational complexity of different algorithms.
Table 2. Computational complexity of different algorithms.
AlgorithmComputational Complexities
MMV-FOCUSS O ( ( N M N d N s L + ( N M ) 3 + 2 ( N M ) 2 N d N s + N M ( N d N s ) 2 ) K )
MMV-FCSBL O ( ( 5 N d N s ( N M ) 2 + ( N M ) 3 + ( 2 N d N s L + 4 N d N s + L ) ( N M ) + 3 N d N s + L ) K )
MMV-ADMM O ( ( 2 N M N d N s L + ( N d N s ) 2 L + 3 N M L + N d N s L ) K )
2DMA-Net O ( ( 2 N d N s L + N M ( N d + N s ) L + ( N + M ) N d N s L ) K )
Generator GYX O ( e = 1 E c e f e 2 n e N d N s ) = O ( 27516 N d N s )
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zou, B.; Wang, X.; Feng, W.; Zhu, H.; Lu, F. DU-CG-STAP Method Based on Sparse Recovery and Unsupervised Learning for Airborne Radar Clutter Suppression. Remote Sens. 2022, 14, 3472. https://doi.org/10.3390/rs14143472

AMA Style

Zou B, Wang X, Feng W, Zhu H, Lu F. DU-CG-STAP Method Based on Sparse Recovery and Unsupervised Learning for Airborne Radar Clutter Suppression. Remote Sensing. 2022; 14(14):3472. https://doi.org/10.3390/rs14143472

Chicago/Turabian Style

Zou, Bo, Xin Wang, Weike Feng, Hangui Zhu, and Fuyu Lu. 2022. "DU-CG-STAP Method Based on Sparse Recovery and Unsupervised Learning for Airborne Radar Clutter Suppression" Remote Sensing 14, no. 14: 3472. https://doi.org/10.3390/rs14143472

APA Style

Zou, B., Wang, X., Feng, W., Zhu, H., & Lu, F. (2022). DU-CG-STAP Method Based on Sparse Recovery and Unsupervised Learning for Airborne Radar Clutter Suppression. Remote Sensing, 14(14), 3472. https://doi.org/10.3390/rs14143472

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop