Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Temperature and Frequency Dependence of Human Cerebrospinal Fluid Dielectric Parameters
Previous Article in Journal
Special Issue: Artificial Intelligence and Smart Sensor-Based Industrial Advanced Technology
Previous Article in Special Issue
Uncertainty-Aware Depth Network for Visual Inertial Odometry of Mobile Robots
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Direction-of-Arrival Estimation with Multi-Task Learning

1
Department of Informatics, Systems and Communication, University of Milano-Bicocca, 20126 Milan, Italy
2
MBDA Missile Systems, 80070 Fusaro, Italy
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(22), 7390; https://doi.org/10.3390/s24227390
Submission received: 23 September 2024 / Revised: 11 November 2024 / Accepted: 19 November 2024 / Published: 20 November 2024
Figure 1
<p>Uniform Linear Array (ULA); <span class="html-italic">d</span> is the distance between the sensors; <math display="inline"><semantics> <msub> <mi>θ</mi> <mi>i</mi> </msub> </semantics></math> is the angle of arrival of the impinging signal, and <span class="html-italic">M</span> is the number of sensor array antennas.</p> ">
Figure 2
<p>Architecture of the proposed multi-task CNN for DOA estimation. The network processes the signal covariance matrix through the backbone. The resulting feature vector is passed to two branches: the Number-of-Source estimator predicts the Number of Sources <math display="inline"><semantics> <mi mathvariant="bold-italic">b</mi> </semantics></math> (i.e., a binarized version of the logits <math display="inline"><semantics> <mi mathvariant="bold-italic">s</mi> </semantics></math>); the Direction-of-Arrival estimator provides multiple angles of arrival, denoted as <math display="inline"><semantics> <mi mathvariant="bold-italic">d</mi> </semantics></math>, corresponding to the number of angles <math display="inline"><semantics> <mi mathvariant="bold-italic">b</mi> </semantics></math> predicted by the other branch. A compound loss <math display="inline"><semantics> <mi mathvariant="script">L</mi> </semantics></math> is used to optimize the model based on the two task-specific losses.</p> ">
Figure 3
<p>Ball chart reporting the RMSE versus accuracy. The size of each ball corresponds to the number of model parameters.</p> ">
Figure 4
<p>The DOA estimation performance for the T1 test set at varying SNRs divided by (<b>a</b>) one signal only, (<b>b</b>) two signals, and (<b>c</b>) three signals.</p> ">
Figure 5
<p>Boxplots showing CRLB index distributions for (<b>a</b>) test sets T1, T3, and T4, (<b>b</b>) various snapshots within T2, and (<b>c</b>) various snapshots within T5.</p> ">
Figure 6
<p>Scenario-independent total performance comparison.</p> ">
Versions Notes

Abstract

:
There are numerous methods in the literature for Direction-of-Arrival (DOA) estimation, including both classical and machine learning-based approaches that jointly estimate the Number of Sources (NOS) and DOA. However, most of these methods do not fully leverage the potential synergies between these two tasks, which could yield valuable shared information. To address this limitation, in this article, we present a multi-task Convolutional Neural Network (CNN) capable of simultaneously estimating both the NOS and the DOA of the signal. Through experiments on simulated data, we demonstrate that our proposed model surpasses the performance of state-of-the-art methods, especially in challenging environments characterized by high noise levels and dynamic conditions.

1. Introduction

Direction-of-Arrival (DOA) estimation is a task that finds applications in various fields, including acoustics, wireless communications, sonar, and radar [1,2,3,4]. To address this problem, several methods have been proposed that can be grouped into traditional methods [1,5,6] and learning-based methods, which are mainly based on deep learning models [7,8]. DL methods have shown impressive advantages when the amount of data used for training is sufficiently large and, most importantly, if the data used for training are representative of the reference scenario [9]. From a strictly technical point of view, DOA estimation is a regression problem where, in the most general case, the dimension of the target is not known a priori. This makes it challenging to apply traditional machine learning methods, which typically require training a model for each source. This issue can be encountered in approaches that use models such as Support Vector Regressors (SVRs) [10] or Support Vector Machine (SVM) [11]. There are also strategies that employ Radial Basis Functions (RBFs) [12,13]. In the latter case, the Number of Sources (NOS) is fixed and assumed to be known a priori, while in the former, which considers the 2D scenario, the problem is addressed by dividing the DOA grid into spatial sub-regions.
Previous methods have relatively good performance in situations where the noise power is very low, but the quality of the estimates degrades as the Signal-to-Noise Ratio (SNR) decreases and approaches zero (i.e., when the noise power is similar to that of the signal). Recently, several solutions have been proposed that utilize Deep Neural Networks (DNNs) [14]. These solutions employ various types of architectures, including Fully Connected Neural Networks (FCNNs) [9,15,16], Convolutional Neural Networks (CNNs) [17,18,19,20,21], and Recurrent Neural Networks (RNNs) [8,22]. The different approaches not only differ in the type of architecture, but also in the input used. Some solutions employ the covariance matrix (separating real and imaginary parts), sometimes vectorized [23,24], or distributed over multiple channels using CNNs [18,25], while others utilize complex-valued spatiotemporal neural networks [7]. Conversely, some methods directly use the received signal or operate in the frequency domain [26,27]. Several methods combine the source detection phase with the DOA determination phase or propose separate solutions for performing the two tasks. Some of these methods assume the number of signals to be known a priori [15,25]. In this regard, some neural models have been presented as replacements for traditional model order estimation methods for identifying the NOS [28,29]. In [28], two distinct types of input are evaluated: the real-valued vector consisting of the real and imaginary parts of the signal, and the covariance matrix of the received signals. In [29], the eigenvalues of the received signal’s covariance matrix are used as input for a regression network (ERNet) and for a classification network (ECNet). A similar strategy can be found in [26], where a neural regression model is implemented with multiple output branches, one of which is specialized in identifying the NOS. The adoption of deep learning-based approaches enhances performance and provides scalable, adaptable solutions that seamlessly integrate with other DOA-related tasks like AESA beamforming [30].
In this paper, we present a novel CNN architecture designed to receive the covariance matrix of the source waveform as input and perform simultaneous estimation of the NOS and DOA. Our approach treats DOA estimation as a regression problem, while NOS estimation leverages an ordinal regression technique. The use of ordinal regression overcomes the limitation of previous methods by considering the relative distance between an erroneously predicted data sample and its ground-truth label. The model is trained end-to-end in a multi-task manner, enhancing its ability to utilize shared representations. This is especially useful in complex settings where traditional methods falter due to variable signal patterns, as it allows the model to understand and integrate the interactions and dependencies between the NOS and DOA tasks. To thoroughly evaluate the robustness and effectiveness of our proposed method, we conduct a comprehensive series of experiments. These tests assess the model’s performance on simulated data with characteristics different from those in the training dataset.
The rest of this paper is as follows. A review of DOA methods is given in Section 2. Section 3 describes the signal model. Section 4 details the proposed multi-task CNN for DOA estimation. Section 5 and Section 6 present simulation results and conclusions, respectively.

2. Related Work

The beamforming method, a fundamental technique for DOA estimation, extends the Discrete Fourier Transform (DFT) from the time domain to the spatial domain. However, it becomes impractical when the beamwidth exceeds the spacing between the sources [1]. Following the development of beamforming, various methods based on spatial spectral estimation and maximum likelihood [31] were introduced for DOA estimation. Notable examples include the MUltiple SIgnal Classification (MUSIC) algorithm [5], the Estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT) algorithm [32], and their variations. The MUSIC algorithm is widely used for DOA estimation, constructing spatial spectral peaks by separating the signal and noise subspaces, which are orthogonal. The DOA can then be estimated from the corresponding spectral peaks using this orthogonal property. The CS-MUSIC algorithm [33], a variant of MUSIC, combines hybrid parametric minimization with MUSIC to address coherence limitations in DOA estimation. ESPRIT, another popular method, splits the array into two identical subarrays with corresponding elements paired at equal distances. This structure ensures that the incident angle differs by only one rotationally invariant factor between the two subarrays, allowing the DOA to be determined by solving for a generalized eigenvalue. Unitary ESPRIT [34], a variant of ESPRIT, exploits the centrosymmetric property of isometric linear matrices, transforming complex covariance matrices into real symmetric ones. This reduces computational complexity and enhances accuracy in DOA estimation. Extended-Phase Interferometry (EPI) combines phase and amplitude data to improve robustness in noisy environments. Using extended least-squares, EPI enhances accuracy over traditional phase interferometry while keeping computational demands moderate [35]. In [36], a low-power AoA estimation system has been proposed for IoT networks like LoRa. Phase data are sent to the cloud for processing, enabling a lightweight, cost-effective design suited for power-constrained applications with limited local processing. A fully digital, reconfigurable FPGA-based architecture for AoA estimation using phase interferometry, focused on low-latency real-time processing, has been proposed in [37]. Its modular design minimizes computational delays, making it ideal for applications needing rapid, precise updates, especially in environments benefiting from hardware-based signal processing.
Another class of DOA estimation techniques relies on sparse representation. In [6], a sparse representation-based method was proposed, using sensor measurements to promote sparsity and improve estimation accuracy via the l 1 -norm penalty. In [38], compressive sensing was applied to DOA estimation, where the method estimates DOA by reconstructing the sparse signal vector [39]. A gridless DOA estimation approach based on a covariance fitting criterion was introduced in [40], offering continuous parameter estimation over time. This method avoids the need for partitioning the angular space into small grids, thus eliminating discretization issues.
In recent years, deep learning [41] has been applied to radio signal processing tasks, including signal detection [42,43], modulation recognition [44,45,46], channel estimation [47], and information recovery [48]. Similarly, deep learning-based methods have emerged as promising alternatives to traditional Direction-of-Arrival (DOA) estimation techniques [24,49]. Many of these methods approach the problem through classification models [17,18,24,49,50,51]. In [24], the authors tackled DOA estimation using a Deep Neural Network (DNN) with the lower triangular part of the covariance matrix as input for two signal sources. The network was trained in parallel on two different datasets: one with randomly generated signal angles and another with signal angles restricted to 1° intervals. A prediction was considered successful if at least one of the two DNNs accurately estimated the DOAs. However, no standardized procedure for selecting the appropriate DNN architecture for this problem was proposed. In [49], another DNN was used for estimating the DOAs of two sources, with four intermediate layers and nine possible choices for the number of units in each. A total of 36 DNNs were trained, and the one yielding the highest accuracy was chosen for DOA estimation. In [17], a deep convolutional network was employed to estimate DOA under sparse prior conditions. This approach reconstructed the spatial spectrum by analyzing the covariance matrix data, and a Convolutional Neural Network (CNN) was trained to extract spectral features. Different activation functions were tested in the hidden layers to optimize performance, while spatial sparsity was leveraged to improve accuracy compared to earlier methods. The design and training of the network were further enhanced by integrating prior information. The authors of [50] introduced CNN-based DeepMUSIC, a deep learning framework that divides the DOA region into sub-regions, with each sub-region handled by a specialized CNN for more precise DOA estimation. In [21], DOA is formulated as a multi-label classification problem, where the angular space is discretized into a grid of multiple classes, and a CNN is used to estimate the angles of arrival of the incoming source signals. In [18], the challenge of DOA estimation in low-Signal-to-Noise Ratio (SNR) conditions was addressed by using multichannel data as input, representing the real part, imaginary part, and phase of the covariance matrix as three separate channels. A 2D convolutional layer was used to extract meaningful features for more accurate DOA estimation. In [51], an online DOA estimation method was proposed by combining CNN with Long Short-Term Memory (LSTM) networks. This hybrid approach enhanced DOA estimation by capturing temporal dependencies in the data.
Deep learning-based DOA estimation methods can also be designed by combining encoders with neural networks. In [9], a method was introduced using a combination of self-encoder and Fully Connected Neural Networks, specifically a Multilayer Perceptron (MLP). This approach involved using a multilayer self-encoder to divide the signal into spatial sub-regions, similar to a spatial filter, and employing a classifier for spatial spectrum estimation in each sub-region. The DNN model was trained with array-specific outputs, making it robust against array imperfections. In [52], a model based on residual networks (ResNet) was proposed to address array defects. The model detected signals within sub-regions through a spatial classification network, followed by classification of outputs from different sub-regions using ResNet, leading to improved accuracy in DOA estimation.
Finally, in [53], a compressed data-based neural network model was introduced, which used joint training of an encoder and classifier for accurate DOA estimation. This model took encoded data from each subarray as input to the classifier, enabling accurate estimation even with limited data. To further simplify the process and avoid computing the covariance matrix, Zheng et al. [54] proposed directly using the raw in-phase and quadrature components of the signal as input to a DNN for DOA estimation.

3. Signal Model

Consider a Uniform Linear Array (ULA) composed of M physical sensors spaced at a distance d from each other [55] as shown in Figure 1.
Assuming a scenario where there are K coplanar far-field narrowband signals, s ( t ) = s 1 , s 2 , , s K , coming from distinct directions θ = θ 1 , θ 2 , , θ K impinging on the array. The signal received by the array at snapshot t is expressed as
x Ω ( t ) = A Ω ( θ ) s ( t ) + n Ω ( t ) , t = 1 , 2 , , T ,
where Ω = { Ω 1 , Ω 2 , , Ω M } denotes the index set of array sensors with Ω m being the position of the m-th sensor with ordered rank Ω 1 < Ω 2 < < Ω M . s ( t ) R K × 1 and n Ω ( t ) R M × 1 denote the transmitted signal and the additive Gaussian sensor noise at the snapshot t, respectively. Let A Ω ( θ ) = [ a Ω ( θ 1 ) , a Ω ( θ 2 ) , , a Ω ( θ K ) ] be the M × K manifold matrix of the signal direction vectors with
a Ω ( θ ) = [ 1 , e j 2 π d λ sin θ , e j 2 2 π d λ sin θ , , e j ( M 1 ) 2 π d λ sin θ ] ,
being the M × 1 steering vector, where j = 1 , λ is the wavelength, and d is the array interval. Therefore, the covariance matrix of the received signal x Ω ( t ) is
R Ω = E [ x Ω ( t ) x Ω H ( t ) ] = A Ω ( θ ) R s s A Ω H ( θ ) + R n n ,
where E ( · ) and ( · ) H , respectively, denote the expectation operator and the Hermitian transpose (conjugate transpose) of a matrix. R s s = E [ s ( t ) s H ( t ) ] denotes the covariance matrix of the source waveform, and R n n = σ 2 I M represents the noise covariance matrix with σ 2 the noise power and I M the M × M identity matrix.

4. The Proposed Method

The proposed model is depicted in Figure 2. It takes a three-dimensional matrix as input, where the spatial dimensions correspond to the covariance matrix, while the two channels represent the real and imaginary parts of the matrix. The model architecture comprises three convolutional blocks, which are shared between two branches, each specialized in distinct tasks. One branch is responsible for identifying the NOS, while the second branch handles the DOA estimation.
In our method, DOA estimation is formulated as a regression problem, while source counting is achieved by using ordinal regression. Specifically, we exploit the COnsistent RAnk Logits (CORAL) framework [56]. The ground-truth NOS K ^ = | y | is first extended into K m a x 1 binary labels s ^ ( 1 ) , , s ^ ( K m a x 1 ) such that s ^ ( k ) { 0 , 1 } indicates whether s ^ ( k ) exceeds rank r k . The predicted rank index is given by
K = 1 + k = 1 K m a x 1 b k ,
where b k { 0 , 1 } is the prediction of the k-th binary classifier in the output layer. The required rank-monotonicity is achieved by using a dense layer that maps features into a single value and K m a x 1 bias parameters.
The CORAL framework suits this application by treating NOS estimation as a ranked prediction problem, where the model distinguishes ordinal categories of source counts. This approach leverages ordinal relationships, capturing the incremental nature of source count estimation and reducing prediction error by considering the distance between predicted and actual counts, enhancing NOS estimation even in challenging cases.

4.1. Preprocessing

The covariance matrix R Ω is unknown and is usually replaced by its sampled estimate R ^ Ω which can be calculated according to the T snapshots. The input data R ¯ Ω of our model are real-valued M × M × 2 tensors, where the third dimension represents different channels. In particular, the first channel is the real part of R ^ Ω :
R ¯ Ω [ : , : , 1 ] = R e { R ^ Ω } ;
the second channel is the imaginary part of R ^ Ω :
R ¯ Ω [ : , : , 2 ] = I m { R ^ Ω } .
Compared to the input of the model in [18], we normalize the input by using a 1 / | | R ¯ Ω | | 2 . This choice is motivated by the fact that, in this way, the model is more resilient to variations in the input’s order of magnitude.

4.2. Architecture

In this section the architecture of the proposed method for DOA estimation is presented. First, the backbone architecture is detailed, then a description of the Number-of-Source estimator and the Direction-of-Arrival estimator is given.

4.2.1. Backbone

The backbone of the proposed model is a shared CNN between the two tasks, which is formalized as follows:
f B ( R ¯ Ω ) = f 3 ( f 2 ( f 1 ( R ¯ Ω ) ) ) = e .
Each function { f i ( · ) } i = 1 , 2 , 3 represents a series of layers: a 2D convolutional layer with 128, 64, and 64 filters, respectively, followed by batch normalization, and a ReLU activation. The output of the last layer of the backbone, e , is flattened and passed to subsequent layers for further processing. A detail of each processing step of the backbone is given in Table 1.

4.2.2. Number-of-Source Estimator

The branch that performs ordinal regression to identify the NOS takes as input the feature vector calculated by the backbone and estimates the NOS, n = h N o S ( e ) . The NOS branch comprises four blocks. The first three blocks consist of a dense layer with 1024, 512, and 256 channels, respectively, followed by a Swish activation function [57]. The branch ends with the CORAL rank predictor layer. This last layer outputs a vector s = { s 1 , s i , , s K m a x } with s i R . The NOS K is obtained by applying Equation (4). The binary predictions, b = { b 1 , b i , , b K m a x } , are obtained via b i = 1 { s i 0.5 } .

4.2.3. Direction-of-Arrival Estimator

The branch that estimates the DOA consists of two dense blocks, similar to those in the ordinal branch but with dropout layers (dropout rate of 0.2), followed by an output dense layer. More in detail, the two dense layers have 1024 and 512 channels, respectively. The last dense layer maps the 512-dimensional feature vector into K m a x real values to form the vector d = { d 1 , d i , , d K m a x } with d i R of the angles of arrival. The final DOA prediction is derived through element-wise multiplication between the binary vector of the NOS, b , and the vector containing the arrival angles, d , as follows d = d b .

4.3. Training Procedure

The proposed model is trained end-to-end for 200 epochs with a batch size of 32 samples by using the Adam optimizer [58], with an initial learning rate of 0.001 , which is reduced by 0.5 every ten epochs. The early stopping technique is implemented to terminate training if the Root Mean Squared Error (RMSE), evaluated on the validation dataset, shows no improvement for 20 consecutive epochs.
Our model is optimized by exploiting the following loss function:
L = λ 1 L M S E + λ 2 L B C E ,
where λ 1 and λ 2 are empirically set to 1.0 and 0.8 , respectively. It consists of the Mean Squared Error (MSE) calculated only on the angles predicted for the first K signals identified by the NOS branch as follows:
L M S E ( d , d ^ ) = 1 N K n = 1 N k = 1 K ( d ^ k ( n ) d k ( n ) ) 2 ,
where N is the number of samples, K is the number of identified signals, and d and d ^ are the predicted and the ground-truth angles, respectively. The second loss of Equation (8) is the Binary Cross-Entropy (BCE), which is defined as follows:
L B C E ( s , s ^ ) = 1 N K n = 1 N k = 1 K m a x 1 log ( s k ( n ) ) s ^ k ( n ) + log ( 1 s k ( n ) ) ( 1 s ^ k ( n ) ) ,
where N is the number of samples, K m a x is the maximum number of signals, s and s ^ are the predicted and the ground-truth logits, respectively.

5. Simulations and Analysis of the Results

5.1. Simulation Settings

The samples used for training and evaluating the models are generated following the data model presented in Section 3.
For the generation of a dataset sample, the number of signals K is first randomly generated from { 1 , , K m a x } , then K angles are randomly sampled from θ , + θ . The covariance matrix R Ω can be then estimated by exploiting the previous values and the fixed parameters of the ULA. The label associated with the covariance matrix is y = θ 1 , , θ K Hence, each sample consists of a pair R Ω , y . The parameters used in our simulations are designed to closely reflect real-world scenarios, incorporating appropriate approximations as recommended in [59]. Specifically, the ULA in this work consists of 10 elements ( M = 10 ) with an inter-element spacing of d = 0.15 , corresponding to a wavelength of λ = 0.3 . Additionally, we set the maximum number of impinging signals to K m a x = 3 . We assume a sampling rate twice the signal frequency, following the Nyquist criterion to prevent aliasing [60]. We use a 32-bit ADC resolution. Finally, multipath components were set to zero, focusing on direct line-of-sight signals to isolate DOA estimation performance under controlled conditions.
We generate a dataset containing data samples with angles in the range 60 , + 59 , affected by variable SNRs uniformly sampled from the intervals [ 13 , 7 ] , [ 3 , + 3 ] , and [ + 7 , + 13 ] dBs. Each sample consists of a total of T = 2000 snapshots.
The generation of data samples begins by generating the entire DOA grid in the range of 60 , + 59 with a step size of 1°. From this grid, all possible combinations of 1, 2, and 3 sources are created. For combinations involving two or three sources, the DOA sequences are randomly permuted. This process yields a total of 288,100 DOA combinations. For each DOA combination, three samples are generated for each one of the three SNR intervals. Considering the DOA grid, the various SNRs and the three sample variants, the dataset comprises 2,592,900 samples. We randomly split this dataset into training and validation sets, allocating 80% for training and 20% for validation. Consequently, the training set consists of 2,074,320 samples, while the validation set contains 518,580 samples.
We then generate seven distinct test sets to assess the method’s performance on unseen data and its capability to generalize to scenarios beyond those simulated in the training set. These test sets incorporate varying conditions—such as phase shifts, modulation types, and carrier frequencies—that impact both the signal waveform and the resulting covariance matrix. Since the covariance matrix captures the correlation structure of signals received across array sensors, these conditions affect the input to the proposed model in unique ways, altering spatial and temporal correlations.
T1.
Same characteristics as the training set
This test set is designed with the same characteristics as the training set.
T2.
Different number of snapshots
This test set contains samples with the same characteristics as the training set in terms of grid of arrival angles and noise levels but varies the number of snapshots T with values { 5 , 100 , 200 , 500 , 1000 } . Testing with different snapshot counts is crucial to assess the model’s performance under varying data availability, from limited snapshots (e.g., T = 5 ) to more extensive datasets (e.g., T = 1000 ), simulating real-world conditions and challenges.
T3.
Various SNRs
This test set consists of samples generated by simulating signals with SNR levels different from those used in the training set, specifically { 20 , 15 , 5 , 5 , 15 , 20 } dB. Testing with unseen SNR levels is crucial to evaluate the model’s generalization ability in real-world scenarios, where signal quality can vary significantly. Low SNR levels (e.g., 20 dB ) represent highly noisy environments, challenging the model’s robustness and noise tolerance, while higher SNR levels (e.g., 20 dB ) simulate clearer signals where precision becomes key.
T4.
Off-grid angles
This test set comprises samples with arrival directions that are not part of the training grid (off-grid angles), specifically [ 59.7 , + 58.3 ] . Experimenting with off-grid signals is essential for evaluating the model’s ability to generalize beyond predefined scenarios and handle real-world conditions where signal arrival angles rarely align perfectly with the training grid.
T5.
Combination of T2, T3 and T4
This test set combines the characteristics of the T2, T3, and T4 test sets, meaning all challenging conditions are present simultaneously. This makes the evaluation data highly realistic, as they reflect real-world scenarios where such conditions often co-occur.
T6.
Various phases
This test set consists of signals with phase shifts in the range [ 45 , 90 , 180 ] , which are not present in the training set or the T1 test, where the phase is fixed at 0 . Testing with varying phase values is essential to assess the model’s robustness to phase variations, a common occurrence in real-world communication and radar systems due to factors like signal propagation and modulation schemes.
T7.
Various frequencies
This test set is designed to evaluate the robustness of the models with respect to changes in signal frequency. The same parameters as in T1 were used, except that the frequency varies. Different carrier frequencies were applied while keeping the sensor positions fixed. Each frequency has its own wavelength and wavenumber, meaning the inter-element spacing is not normalized to each signal’s wavelength. Four subsets were created, each with signals at a different frequency: 2.4 GHz, 3 GHz, 5 GHz, and 10 GHz. These frequencies were selected due to their widespread use across various applications. Higher frequencies (5 GHz and 10 GHz) are typically better suited for short-range, high-bandwidth applications, while lower frequencies (2.4 GHz and 3 GHz) offer greater coverage and signal penetration.
T8.
Modulation types
This test set is similar to T1, except that different modulation types are applied to simulate a variety of signal behaviors. Amplitude Modulation (AM) introduces a variation in the signal’s amplitude over time, creating fluctuations in strength that can be detected and analyzed. Frequency Modulation (FM) produces a signal with a varying instantaneous frequency, altering how rapidly the signal’s phase evolves, which affects its spectral properties. Phase Shift Keying (PSK) changes the signal’s phase according to a specific modulation scheme, typically used to encode information in discrete phase shifts, making it ideal for digital communication.

5.2. Results

The performance of the proposed model for DOA estimation is compared with five state-of-the-art DOA methods that have publicly available implementations. MUltiple SIgnificant Classification (MUSIC) identifies DOAs in correspondence with the pseudo-spectra identified on a predefined grid [5]. Root MUSIC (R-MUSIC) estimates the angular directions from the solutions of higher-order polynomials [61]. The aforementioned methods are covariance-based techniques that demand sufficient data snapshots for accurate DOA estimation and often assume known NOS. To ensure consistency with the experimental setup of other methods where the NOS is unknown, we applied the Akaike Information Criterion (AIC) to estimate the NOS before performing DOA estimation [62]. Multi-snapshot Newtonized Orthogonal Matching Pursuit (MNOMP) uses Newton refinement and feedback strategy for DOA estimation, leveraging Fast Fourier Transform (FFT) to keep computation complexity low [63]. The Multi-Task Autoencoder with Parallel multilayer Classifiers (MTAPC) model effectively estimates spatial spectra in complex scenarios, primarily thanks to the enhanced representation achieved by the multi-task autoencoder [9]. DNNDOA [18] is a CNN designed for DOA estimation in the low-SNR regime as a multi-label classification task, considering an on-grid approach.
The Cramér–Rao Lower Bound (CRLB) provides a theoretical benchmark for the minimum variance of unbiased estimators, making it an essential measure for complementing and validating the results obtained by existing methods in DOA estimation. Comparing DOA methods with the CRLB helps assess their proximity to optimal performance, highlighting both efficiency and unbiasedness. Since the CRLB accounts for factors such as noise and sensor configuration, it serves as a crucial reference for validating results on simulated data under various conditions. Methods that approach the CRLB reach the theoretical performance limit, while those that exceed it reveal areas for improvement [64]. Therefore, we include the CRLB in our comparison of results.
The results are presented using RMSE to evaluate the precision of DOA estimation, while accuracy is employed to estimate the NOS, measured as the fraction of correctly estimated sources. RMSE is calculated using the following equation [26]:
RMSE = 1 K k = 1 K min j ( d k d ^ j ) 2 ,
where d k indicates the predicted angle, d ^ j represents the ground-truth angle, and K is the number of signals. Ground-truth and predicted signals are paired using a nearest-neighbor metric, ensuring that each predicted angle is compared to its closest ground-truth angle, resulting in a more accurate and lower RMSE value. When the number of predicted sources does not match the number of ground-truth sources, special handling is required for RMSE computation. Table 2 provides an example to demonstrate how these cases are addressed.
For test set T1, we report results in Table 3. As it is possible to see, the proposed method achieved the best performance with respect to the competitors. On the other hand, MNOMP achieves the worst RMSE and accuracy. From the perspective of efficiency, we evaluate methods based on inference time, computational complexity, and the number of parameters. In terms of inference time, deep learning methods are noticeably slower than traditional approaches like MUSIC, R-MUSIC, and MNOMP. When measuring computational complexity in Mega-FLOPS (MFLOPS), these traditional methods again prove more efficient than deep learning approaches. Among the deep learning methods, DNNDOA has the highest computational complexity, reaching 2.43 MFLOPS. Despite the fact that our proposed method has significantly more parameters than MUSIC, R-MUSIC, and MTAPC, Figure 3 demonstrates its superior performance in both RMSE and accuracy. Interestingly, DNNDOA, which has a similar parameter count to our model, performs the worst in terms of accuracy. In Figure 4, the RMSE is decomposed by the number of sources and varying SNR. Notably, the RMSE value for three sources, shown in panel (c), is higher than that for one and two sources. Among the algorithms evaluated, DNNDOA exhibits the greatest sensitivity to high SNR values.
In the T2 test set (see Table 4), performance declines with shorter snapshots for all methods. Our method outperforms others, with RMSE and accuracy of 11.81 and 47.04 ( T = 5 ) compared to 0.32 and 99.58 ( T = 1000 ). In the T3 test set, as seen in Table 5, the methods exhibit comparable RMSE performance, with the exception of DNNDOA. MTPAC achieves the lowest error, followed by R-MUSIC and MUSIC, with our method close behind. This suggests that the other methods demonstrate slightly better generalization to noise levels not covered in the training set.
The performance of the T4 test set is presented in Table 5. As before, it clearly demonstrates that the proposed method outperforms the other methods. In a broader context, when comparing these results with those in Table 3, it turns out that off-grid angles have a minimal impact on the performance of the methods under consideration. Analyzing the results presented in Table 6 for the T5 test set, it is evident that the proposed method’s RMSE performance remains relatively stable across varying snapshot lengths. In fact, our method outperforms the competitors in most cases, with the exception of RMSE at T = 1000 . It is worth noting the stark underperformance in estimating the NOS by the MUSIC and R-MUSIC methods when dealing with snapshots of fewer than 500, with rates as low as 0.00%. In Table 7, it achieves the lowest RMSE and high accuracy across all phases, proving robust against phase variability, while other methods like DNNDOA and MNOMP display variable performance with higher RMSE, particularly at 45° and 90° phases. In Table 8, the proposed method excels across the frequency spectrum, especially at higher frequencies (5 GHz and 10 GHz), where it achieves an RMSE of 18.56 and nearly 93% accuracy. While MUSIC and R-MUSIC maintain high accuracy, they show slightly higher RMSE, indicating some limitations at high frequencies. Table 9 shows that the proposed method also leads in handling different modulation types, particularly FM and PSK, achieving low RMSE and high accuracy.
The boxplots in Figure 5 illustrate the CRLB index distributions for the different test sets and snapshot counts. In the first plot (T1, T3, T4), the CRLB values are concentrated between 10 3 and 10 1 degrees squared. T3 shows a larger spread, indicating more variability in estimation error. In the second plot (T2), as the number of snapshots increases, the CRLB decreases significantly, reflecting improved estimation accuracy with more data. At 1000 snapshots, the CRLB is concentrated between 10 2 and 10 3 . The third plot (T5) follows a similar trend, though with higher overall CRLB values compared to T2, indicating more challenging estimation conditions for T5. Overall, more snapshots lead to a lower and narrower CRLB, highlighting improved performance.
By comparing CRLB values with those achieved by the proposed method on the various test sets, it is possible to notice that as snapshot count increases, the RMSE of the proposed DOA estimation method converges toward the CRLB, indicating more efficient performance with larger data samples. At low snapshot counts (5 and 100), the RMSE is notably higher than the CRLB, reflecting increased estimation uncertainty with limited data. However, with more snapshots (200, 500, and especially 1000), the RMSE closely approaches the CRLB, demonstrating the method’s ability to achieve near-optimal accuracy when sufficient data are available. For instance, with 1000 snapshots in T2, the RMSE of 0.32 is close to the CRLB of 0.22 , and in T5, the RMSE of 5.31 is near the CRLB of 5.72 .
To obtain a scenario-independent comparison of each method’s relative performance, we normalize the RMSE of each method by the CRLB for the corresponding scenario. Since the CRLB represents the best possible performance achievable in a given scenario, dividing the squared error by this bound yields a normalized measure, which is independent of the specific scenario’s characteristics. This normalization minimizes the impact of scenario variability on the RMSE, thus reducing the overall uncertainty when averaging performance across scenarios. The result is an effectiveness metric for each method, defined as the ratio of RMSE to CRLB, averaged over all scenarios. As seen in Figure 6, our approach more accurately reflects the performance disparities among estimators, making it evident how each one approaches or deviates from the theoretical CRLB limit.

5.3. Ablation Studies

In this section, we present the results of two ablation studies to evaluate the impact of training the proposed method in stages and varying the weights of the loss functions. In the first experiment, the learning process is divided into two steps. First, only the backbone and NOS branch are trained. Then, in the second step, the DOA branch is trained, leveraging the weights previously learned by the backbone and NOS estimator, without fine-tuning them. This incremental training approach allows us to evaluate the potential disadvantages of two-step training for DOA estimation compared to multi-task training, where shared features are learned simultaneously. In the second experiment, we explore various configurations of the λ 1 and λ 2 weights in Equation (8) to assess their impact on optimization and accuracy. To this end, we present results for two configurations that significantly differ from the optimal one: λ 1 = 0.2 , λ 2 = 1.0 , and λ 1 = 1.0 , λ 2 = 0.4 .
Results from the T1 test set, i.e., data with the same characteristics as the training set, are reported in Table 10. As it is possible to see, the two-step training approach results in a significantly higher RMSE (32.99) compared to the multi-task approaches, suggesting that learning the shared features simultaneously, as in multi-task training, leads to better performance for DOA estimation. The two configurations with different λ 1 and λ 2 weights show substantial differences in RMSE values. The configuration with λ 1 = 0.2 and λ 2 = 1.0 results in a higher RMSE (0.75) compared to the configuration with λ 1 = 1.0 and λ 2 = 0.4 (0.68), indicating that adjusting the balance of loss functions can improve optimization and reduce error. The proposed method (“Our” model version) achieves the best performance, with the lowest RMSE (0.27) and the highest accuracy (99.58%), suggesting that the specific configuration used in the final model optimally balances the training dynamics and yields the most accurate predictions.

6. Conclusions

In this paper, we proposed a deep multi-task CNN for joint NOS and DOA estimation. Compared with methods in the literature, the proposed method considers the estimation of the NOS as ordinal regression and the estimation of DOA as regression. The previous choices demonstrate the effectiveness of the proposed method even under challenging conditions. As future work, we aim to improve the proposed method by incorporating additional signal properties, such as noise statistics, signal coherence, and temporal snapshot information, to enhance the model’s ability to capture complex data relationships and improve DOA estimation accuracy in challenging environments, as seen in classical detection theory methods. While our current tests include varying SNRs, off-grid angles, and snapshots, future experiments will also involve environmental factors such as multipath propagation, receiver sampling ratio, and ADC quantization bits. Additionally, we plan to validate the method using real-world data, where obstacles and reflections are harder to simulate but crucial for practical applications. This will help refine our approach for real-world scenarios and ensure robustness in diverse environments.

Author Contributions

Conceptualization, S.B., L.C., P.C., P.N., G.P. and P.V.; methodology, S.B., L.C., P.C., P.N. and P.V.; software, L.C. and P.C.; validation, S.B., L.C., P.C., P.N., G.P. and P.V.; formal analysis, S.B., L.C., P.C., P.N. and P.V.; investigation, S.B., L.C., P.C., P.N. and P.V.; resources, S.B. and P.N.; data curation, P.C. and P.V.; writing—original draft preparation, S.B., L.C. and P.N.; writing—review and editing, S.B., L.C., P.N. and P.V.; visualization, S.B., L.C. and P.N.; supervision, S.B., P.N. and G.P.; project administration, P.V. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Krim, H.; Viberg, M. Two decades of array signal processing research: The parametric approach. IEEE Signal Process. Mag. 1996, 13, 67–94. [Google Scholar] [CrossRef]
  2. Van Trees, H. Optimum Array Processing: Part IV of Detection, Estimation, and Modulation Theory; John Wiley & Sons: Hoboken, NJ, USA, 2002. [Google Scholar]
  3. Yadav, S.K.; George, N.V. Coarray MUSIC-group delay: High-resolution source localization using non-uniform arrays. IEEE Trans. Veh. Technol. 2021, 70, 9597–9601. [Google Scholar] [CrossRef]
  4. Liao, B.; Zhang, Z.G.; Chan, S.C. DOA estimation and tracking of ULAs with mutual coupling. IEEE Trans. Aerosp. Electron. Syst. 2012, 48, 891–905. [Google Scholar] [CrossRef]
  5. Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef]
  6. Malioutov, D.; Cetin, M.; Willsky, A.S. A sparse signal reconstruction perspective for source localization with sensor arrays. IEEE Trans. Signal Process. 2005, 53, 3010–3022. [Google Scholar] [CrossRef]
  7. Terabayashi, K.; Natsuaki, R.; Hirose, A. Ultrawideband Direction-of-Arrival Estimation Using Complex-Valued Spatiotemporal Neural Networks. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 1727–1732. [Google Scholar] [CrossRef]
  8. Yao, Y.; Lei, H.; He, W. A-CRNN-Based Method for Coherent DOA Estimation with Unknown Source Number. Sensors 2020, 20, 2296. [Google Scholar] [CrossRef]
  9. Liu, Z.M.; Zhang, C.; Philip, S.Y. Direction-of-Arrival Estimation Based on Deep Neural Networks with Robustness to Array Imperfections. IEEE Trans. Antennas Propag. 2018, 66, 7315–7327. [Google Scholar] [CrossRef]
  10. Wu, L.L.; Huang, Z.T. Coherent SVR Learning for Wideband Direction-of-Arrival Estimation. IEEE Signal Process. Lett. 2019, 26, 642–646. [Google Scholar] [CrossRef]
  11. Gao, Y.; Hu, D.; Chen, Y.; Ma, Y. Gridless 1-b DOA Estimation Exploiting SVM Approach. IEEE Commun. Lett. 2017, 21, 2210–2213. [Google Scholar] [CrossRef]
  12. El Zooghby, A.; Christodoulou, C.; Georgiopoulos, M. A neural network-based smart antenna for multiple source tracking. IEEE Trans. Antennas Propag. 2000, 48, 768–776. [Google Scholar] [CrossRef]
  13. Agatonović, M.; Stanković, Z.; Milovanović, B.; Dončov, N. DOA estimation using radial basis function neural networks as uniform circular antenna array signal processor. In Proceedings of the International Conference on Telecommunication in Modern Satellite Cable and Broadcasting Services (TELSIKS), Nis, Serbia, 5–8 October 2011; Volume 2, pp. 544–547. [Google Scholar] [CrossRef]
  14. Ge, S.; Li, K.; Rum, S.N.B.M. Deep learning approach in DOA estimation: A systematic literature review. Mob. Inf. Syst. 2021, 2021, 6392875. [Google Scholar] [CrossRef]
  15. Guo, Y.; Zhang, Z.; Huang, Y.; Zhang, P. DOA Estimation Method Based on Cascaded Neural Network for Two Closely Spaced Sources. IEEE Signal Process. Lett. 2020, 27, 570–574. [Google Scholar] [CrossRef]
  16. Barthelme, A.; Utschick, W. A Machine Learning Approach to DoA Estimation and Model Order Selection for Antenna Arrays with Subarray Sampling. IEEE Trans. Signal Process. 2021, 69, 3075–3087. [Google Scholar] [CrossRef]
  17. Wu, L.; Liu, Z.M.; Huang, Z.T. Deep Convolution Network for Direction of Arrival Estimation with Sparse Prior. IEEE Signal Process. Lett. 2019, 26, 1688–1692. [Google Scholar] [CrossRef]
  18. Papageorgiou, G.; Sellathurai, M.; Eldar, Y. Deep Networks for Direction-of-Arrival Estimation in Low SNR. IEEE Trans. Signal Process. 2021, 69, 3714–3729. [Google Scholar] [CrossRef]
  19. Chakrabarty, S.; Habets, E.A.P. Multi-Speaker DOA Estimation Using Deep Convolutional Networks Trained with Noise Signals. IEEE J. Sel. Top. Signal Process. 2019, 13, 8–21. [Google Scholar] [CrossRef]
  20. Wu, X.; Yang, X.; Jia, X.; Tian, F. A gridless DOA estimation method based on convolutional neural network with Toeplitz prior. IEEE Signal Process. Lett. 2022, 29, 1247–1251. [Google Scholar] [CrossRef]
  21. Kokkinis, G.; Zaharis, Z.D.; Lazaridis, P.I.; Kantartzis, N.V. Direction of arrival estimation applied to antenna arrays using convolutional neural networks. In Proceedings of the URSI Atlantic and Asia Pacific Radio Science Meeting (AT-AP-RASC), Gran Canaria, Spain, 29 May–3 June 2022; pp. 1–4. [Google Scholar]
  22. Xiang, H.; Chen, B.; Yang, M.; Xu, S.; Li, Z. Improved direction-of-arrival estimation method based on LSTM neural networks with robustness to array imperfections. Appl. Intell. 2021, 51, 1–14. [Google Scholar] [CrossRef]
  23. Ünlerşen, M.F.; Yaldiz, E. Direction of Arrival Estimation by Using Artificial Neural Networks. In Proceedings of the European Modelling Symposium (EMS), Pisa, Italy, 28–30 November 2016; pp. 242–245. [Google Scholar] [CrossRef]
  24. Kase, Y.; Nishimura, T.; Ohgane, T.; Ogawa, Y.; Kitayama, D.; Kishiyama, Y. DOA estimation of two targets with deep learning. In Proceedings of the Workshop on Positioning, Navigation and Communications (WPNC), Bremen, Germany, 25–26 October 2018; pp. 1–5. [Google Scholar] [CrossRef]
  25. Zhu, W.; Zhang, M.; Li, P.; Wu, C. Two-Dimensional DOA Estimation via Deep Ensemble Learning. IEEE Access 2020, 8, 124544–124552. [Google Scholar] [CrossRef]
  26. Bialer, O.; Garnett, N.; Tirer, T. Performance Advantages of Deep Neural Networks for Angle of Arrival Estimation. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3907–3911. [Google Scholar]
  27. Izacard, G.; Bernstein, B.; Fernandez-Granda, C. A learning-based framework for line-spectra super-resolution. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 3632–3636. [Google Scholar]
  28. Barthelme, A.; Wiesmayr, R.; Utschick, W. Model Order Selection in DoA Scenarios via Cross-entropy Based Machine Learning Techniques. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 4–9 May 2020; pp. 4622–4626. [Google Scholar]
  29. Yang, Y.; Gao, F.; Qian, C.; Liao, G. Model-Aided Deep Neural Network for Source Number Detection. IEEE Signal Process. Lett. 2020, 27, 91–95. [Google Scholar] [CrossRef]
  30. Bianco, S.; Napoletano, P.; Raimondi, A.; Feo, M.; Petraglia, G.; Vinetti, P. AESA adaptive beamforming using deep learning. In Proceedings of the IEEE Radar Conference (RadarConf20), Florence, Italy, 21–25 September 2020; pp. 1–6. [Google Scholar]
  31. Miller, M.I.; Fuhrmann, D.R. Maximum-likelihood narrow-band direction finding and the EM algorithm. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1560–1577. [Google Scholar] [CrossRef]
  32. Roy, R.; Kailath, T. ESPRIT-estimation of signal parameters via rotational invariance techniques. IEEE Trans. Acoust. Speech Signal Process. 1989, 37, 984–995. [Google Scholar] [CrossRef]
  33. Parian, M.A.; Ghofrani, S. Using l1, 2 mixed-norm MUSIC based on compressive sampling for direction of arrival estimation. In Proceedings of the International Symposium on Signal Processing and Information Technology (ISSPIT), Abu Dhabi, United Arab Emirates, 7–10 December 2015; pp. 258–263. [Google Scholar]
  34. Haardt, M.; Ali-Hackl, M.E. Unitary esprit: How to exploit additional information inherent in the relational invariance structure. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, Adelaide, SA, Australia, 19–22 April 1994; Volume 4, p. IV-229. [Google Scholar]
  35. Wu, Y.W.; Rhodes, S.; Satorius, E.H. Direction of arrival estimation via extended phase interferometry. IEEE Trans. Aerosp. Electron. Syst. 1995, 31, 375–381. [Google Scholar]
  36. BniLam, N.; Nasser, S.; Weyn, M. Angle of arrival estimation system for LoRa technology based on phase detectors. In Proceedings of the 2022 16th European Conference on Antennas and Propagation (EuCAP), Madrid, Spain, 27 March–1 April 2022; pp. 1–5. [Google Scholar]
  37. Florio, A.; Avitabile, G.; Talarico, C.; Coviello, G. A Reconfigurable Full-Digital Architecture for Angle of Arrival Estimation. IEEE Trans. Circuits Syst. I Regul. Pap. 2023, 71, 1443–1455. [Google Scholar] [CrossRef]
  38. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  39. Massa, A.; Rocca, P.; Oliveri, G. Compressive sensing in electromagnetics-a review. IEEE Antennas Propag. Mag. 2015, 57, 224–238. [Google Scholar] [CrossRef]
  40. Yang, Z.; Xie, L.; Zhang, C. A discretization-free sparse and parametric approach for linear array signal processing. IEEE Trans. Signal Process. 2014, 62, 4959–4973. [Google Scholar] [CrossRef]
  41. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  42. Wei, F.; Zheng, S.; Zhou, X.; Zhang, L.; Lou, C.; Zhao, Z.; Yang, X. Detection of direct sequence spread spectrum signals based on deep learning. IEEE Trans. Cogn. Commun. Netw. 2022, 8, 1399–1410. [Google Scholar] [CrossRef]
  43. Zheng, S.; Chen, S.; Qi, P.; Zhou, H.; Yang, X. Spectrum sensing based on deep learning classification for cognitive radios. China Commun. 2020, 17, 138–148. [Google Scholar] [CrossRef]
  44. Zheng, S.; Zhou, X.; Zhang, L.; Qi, P.; Qiu, K.; Zhu, J.; Yang, X. Toward next-generation signal intelligence: A hybrid knowledge and data-driven deep learning framework for radio signal classification. IEEE Trans. Cogn. Commun. Netw. 2023, 9, 564–579. [Google Scholar] [CrossRef]
  45. Liu, R.; Ma, Y.; Zhang, X.; Gao, Y. Deep learning-based spectrum sensing in space-air-ground integrated networks. J. Commun. Inf. Netw. 2021, 6, 82–90. [Google Scholar] [CrossRef]
  46. Peng, S.; Sun, S.; Yao, Y.D. A survey of modulation classification using deep learning: Signal representation and data preprocessing. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 7020–7038. [Google Scholar] [CrossRef] [PubMed]
  47. Xu, M.; Zhang, S.; Ma, J.; Dobre, O.A. Deep learning-based time-varying channel estimation for RIS assisted communication. IEEE Commun. Lett. 2021, 26, 94–98. [Google Scholar] [CrossRef]
  48. Zheng, S.; Chen, S.; Yang, X. DeepReceiver: A deep learning-based intelligent receiver for wireless communications in the physical layer. IEEE Trans. Cogn. Commun. Netw. 2020, 7, 5–20. [Google Scholar] [CrossRef]
  49. Kase, Y.; Nishimura, T.; Ohgane, T.; Ogawa, Y.; Kitayama, D.; Kishiyama, Y. Performance analysis of DOA estimation of two targets using deep learning. In Proceedings of the 2019 22nd International Symposium on Wireless Personal Multimedia Communications (WPMC), Lisbon, Portugal, 24–27 November 2019; pp. 1–6. [Google Scholar]
  50. Elbir, A.M. DeepMUSIC: Multiple signal classification via deep learning. IEEE Sens. Lett. 2020, 4, 1–4. [Google Scholar] [CrossRef]
  51. Li, Q.; Zhang, X.; Li, H. Online direction of arrival estimation based on deep learning. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2616–2620. [Google Scholar]
  52. Li, J.; Shao, X.; Li, J.; Ge, L. Direction of arrival estimation of array defects based on deep neural network. Circ. Syst. Signal Process. 2022, 41, 4906–4927. [Google Scholar] [CrossRef]
  53. Pavel, S.R.; Zhang, Y.D. Neural network-based compression framework for DoA estimation exploiting distributed array. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 4943–4947. [Google Scholar]
  54. Zheng, S.; Yang, Z.; Shen, W.; Zhang, L.; Zhu, J.; Zhao, Z.; Yang, X. Deep learning-based DOA estimation. IEEE Trans. Cogn. Commun. Netw. 2024, 10, 819–835. [Google Scholar] [CrossRef]
  55. Balanis, C.A. Antenna theory: A review. Proc. IEEE 1992, 80, 7–23. [Google Scholar] [CrossRef]
  56. Cao, W.; Mirjalili, V.; Raschka, S. Rank consistent ordinal regression for neural networks with application to age estimation. Pattern Recognit. Lett. 2020, 140, 325–331. [Google Scholar] [CrossRef]
  57. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  58. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar]
  59. Molaei, A.M.; Zakeri, B.; Andargoli, S.M.H.; Abbasi, M.A.B.; Fusco, V.; Yurduseven, O. A Comprehensive Review of Direction-of-Arrival Estimation and Localization Approaches in Mixed-Field Sources Scenario. IEEE Access 2024, 12, 65883–65918. [Google Scholar] [CrossRef]
  60. Nyquist, H. Certain factors affecting telegraph speed. Trans. Am. Inst. Electr. Eng. 1924, 43, 412–422. [Google Scholar] [CrossRef]
  61. Barabell, A. Improving the resolution performance of eigenstructure-based direction-finding algorithms. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Boston, MA, USA, 14–16 April 1983; Volume 8, pp. 336–339. [Google Scholar]
  62. Akaike, H. Information theory and an extension of the maximum likelihood principle. In Selected Papers of Hirotugu Akaike; Springer: New York, NY, USA, 1998; pp. 199–213. [Google Scholar]
  63. Zhu, J.; Han, L.; Blum, R.S.; Xu, Z. Multi-snapshot Newtonized orthogonal matching pursuit for line spectrum estimation with multiple measurement vectors. Signal Process. 2019, 165, 175–185. [Google Scholar] [CrossRef]
  64. Stoica, P.; Nehorai, A. Performance study of conditional and unconditional direction-of-arrival estimation. IEEE Trans. Acoust. Speech Signal Process. 1990, 38, 1783–1795. [Google Scholar] [CrossRef]
Figure 1. Uniform Linear Array (ULA); d is the distance between the sensors; θ i is the angle of arrival of the impinging signal, and M is the number of sensor array antennas.
Figure 1. Uniform Linear Array (ULA); d is the distance between the sensors; θ i is the angle of arrival of the impinging signal, and M is the number of sensor array antennas.
Sensors 24 07390 g001
Figure 2. Architecture of the proposed multi-task CNN for DOA estimation. The network processes the signal covariance matrix through the backbone. The resulting feature vector is passed to two branches: the Number-of-Source estimator predicts the Number of Sources b (i.e., a binarized version of the logits s ); the Direction-of-Arrival estimator provides multiple angles of arrival, denoted as d , corresponding to the number of angles b predicted by the other branch. A compound loss L is used to optimize the model based on the two task-specific losses.
Figure 2. Architecture of the proposed multi-task CNN for DOA estimation. The network processes the signal covariance matrix through the backbone. The resulting feature vector is passed to two branches: the Number-of-Source estimator predicts the Number of Sources b (i.e., a binarized version of the logits s ); the Direction-of-Arrival estimator provides multiple angles of arrival, denoted as d , corresponding to the number of angles b predicted by the other branch. A compound loss L is used to optimize the model based on the two task-specific losses.
Sensors 24 07390 g002
Figure 3. Ball chart reporting the RMSE versus accuracy. The size of each ball corresponds to the number of model parameters.
Figure 3. Ball chart reporting the RMSE versus accuracy. The size of each ball corresponds to the number of model parameters.
Sensors 24 07390 g003
Figure 4. The DOA estimation performance for the T1 test set at varying SNRs divided by (a) one signal only, (b) two signals, and (c) three signals.
Figure 4. The DOA estimation performance for the T1 test set at varying SNRs divided by (a) one signal only, (b) two signals, and (c) three signals.
Sensors 24 07390 g004
Figure 5. Boxplots showing CRLB index distributions for (a) test sets T1, T3, and T4, (b) various snapshots within T2, and (c) various snapshots within T5.
Figure 5. Boxplots showing CRLB index distributions for (a) test sets T1, T3, and T4, (b) various snapshots within T2, and (c) various snapshots within T5.
Sensors 24 07390 g005
Figure 6. Scenario-independent total performance comparison.
Figure 6. Scenario-independent total performance comparison.
Sensors 24 07390 g006
Table 1. The backbone architecture proposed in this work. Each row describes an operation by specifying the input shape, operator type, number of output channels (c), stride (s), and output shape. A dash “-” indicates that the operator does not require a value for the parameter.
Table 1. The backbone architecture proposed in this work. Each row describes an operation by specifying the input shape, operator type, number of output channels (c), stride (s), and output shape. A dash “-” indicates that the operator does not require a value for the parameter.
InputOperatorcsOutput
2 × M × M Conv2d1282 128 × M / 2 × M / 2
128 × M / 2 × M / 2 BatchNorm2d-- 128 × M / 2 × M / 2
128 × M / 2 × M / 2 ReLU-- 128 × M / 2 × M / 2
128 × M / 2 × M / 2 Conv2d641 64 × M / 2 × M / 2
64 × M / 2 × M / 2 BatchNorm2d-- 64 × M / 2 × M / 2
64 × M / 2 × M / 2 ReLU-- 64 × M / 2 × M / 2
64 × M / 2 × M / 2 Conv2d641 64 × M / 2 × M / 2
64 × M / 2 × M / 2 BatchNorm2d-- 64 × M / 2 × M / 2
64 × M / 2 × M / 2 ReLU-- 64 × M / 2 × M / 2
64 × M / 2 × M / 2 Flatten-- 64 M / 2 M / 2
Table 2. Samples illustrating how RMSE is computed when the number of predicted sources differs from the number of ground-truth sources. In these examples, d ^ represents the ground-truth DOA, and d represents the predicted DOA.
Table 2. Samples illustrating how RMSE is computed when the number of predicted sources differs from the number of ground-truth sources. In these examples, d ^ represents the ground-truth DOA, and d represents the predicted DOA.
CaseDirectionRMSE
| d ^ |   <   | d | d ^ = { 15 , 19 }
d = { 20 , 15 , 32 }
( 15 15 ) 2 + ( 19 + 20 ) 2 2 = 0.71
| d ^ |   >   | d | d ^ = { 19 , 42 , 15 }
d = { 20 , 15 }
( 19 + 20 ) 2 + ( 42 15 ) 2 + ( 15 15 ) 2 3 = 15.60
Table 3. Results reported in terms of RMSE, accuracy (%), inference time (ms) and computational complexity (MFlops) for the test set with the same configuration as the training set. The best result for each metric is shown in boldface.
Table 3. Results reported in terms of RMSE, accuracy (%), inference time (ms) and computational complexity (MFlops) for the test set with the same configuration as the training set. The best result for each metric is shown in boldface.
MethodRMSE (↓)Accuracy (↑)Inference
Time (↓)
Computational
Complexity (↓)
DNNDOA [18]6.56 ± 15.5564.1040.462.43
MNOMP [63]13.76 ± 18.3953.371.920.01
MTAPC [9]1.91 ± 7.3673.8740.271.61
MUSIC [5]1.32 ± 2.5691.201.450.02
R-MUSIC [61]0.60 ± 1.3793.450.740.02
Our0.27± 0.2199.5840.221.11
Table 4. RMSE and accuracy for different T snapshots. The best result for each metric is shown in boldface.
Table 4. RMSE and accuracy for different T snapshots. The best result for each metric is shown in boldface.
Method51002005001000
RMSE (↓)Acc. (↑)RMSE (↓)Acc. (↑)RMSE (↓)Acc. (↑)RMSE (↓)Acc. (↑)RMSE (↓)Acc. (↑)
DNNDOA [18]30.71 ± 17.326.4114.65 ± 19.5340.6411.17 ± 18.2849.608.07 ± 16.6158.307.05 ± 15.9261.87
MNOMP [63]48.56 ± 25.256.7638.05 ± 24.5442.2736.07 ± 24.3055.4035.92 ± 23.9359.7031.73 ± 22.1866.31
MTAPC [9]16.19 ± 10.0232.512.40 ± 7.4963.232.23 ± 7.5368.762.07 ± 7.4972.451.98 ± 7.4273.44
MUSIC [5]19.42 ± 20.852.013.05 ± 1.270.001.73 ± 0.960.010.61 ± 1.7411.751.30 ± 1.6685.34
R-MUSIC [61]19.72 ± 21.291.033.12 ± 2.100.001.79 ± 0.860.010.56 ± 0.5111.751.25 ± 1.4285.45
Our11.81 ± 9.5947.042.24 ± 3.2991.461.13 ± 0.8697.570.47 ± 0.4599.460.32 ± 0.3199.58
Table 5. RMSE and accuracy for various SNRs (i.e., −20, −15, −5, 5, 15, 20) and off-grid angles (i.e., between 59.7 and + 58.3 ). The best result for each metric is shown in boldface.
Table 5. RMSE and accuracy for various SNRs (i.e., −20, −15, −5, 5, 15, 20) and off-grid angles (i.e., between 59.7 and + 58.3 ). The best result for each metric is shown in boldface.
MethodVarious SNRsOff-Grid Angles
RMSE (↓)Accuracy (↑)RMSE (↓)Accuracy (↑)
DNNDOA [18]29.82 ± 15.464.528.31 ± 16.9051.17
MNOMP [63]34.69 ± 31.1612.629.98 ± 7.5156.20
MTAPC [9]2.24 ± 7.1861.521.72 ± 6.3973.66
MUSIC [5]6.02 ± 14.5574.670.34 ± 0.6890.81
R-MUSIC [61]6.02 ± 14.6274.730.32 ± 0.3992.34
Our8.47 ± 14.0487.770.26 ± 0.2399.57
Table 6. RMSE and accuracy for the test set with different T snapshots, off-grid angles and various SNRs. The best result for each metric is shown in boldface.
Table 6. RMSE and accuracy for the test set with different T snapshots, off-grid angles and various SNRs. The best result for each metric is shown in boldface.
Method51002005001000
RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑)
DNNDOA [18]32.60 ± 14.311.9429.91 ± 15.464.3529.68 ± 15.314.4629.60 ± 15.274.4729.59 ± 15.264.46
MNOMP [63]46.28 ± 25.249.0232.54 ± 23.7037.0730.84 ± 23.3344.9330.18 ± 23.2848.4528.38 ± 16.7554.19
MTAPC [9]17.16 ± 12.5331.8213.61 ± 11.7050.6013.08 ± 11.1551.9012.54 ± 10.6154.8212.23 ± 10.3757.95
MUSIC [5]20.34 ± 21.471.0318.83 ± 18.440.0018.48 ± 16.170.0020.02 ± 15.9814.9315.76 ± 14.1170.26
R-MUSIC [61]20.91 ± 18.390.0218.71 ± 16.020.0018.44 ± 16.950.0020.31 ± 15.8414.6915.81 ± 14.4370.32
Our12.20 ± 9.7844.688.74 ± 13.9065.788.39 ± 13.2366.237.90 ± 12.7376.696.31 ± 11.6185.44
Table 7. RMSE and accuracy for various phases of test set T6. The best result for each metric is shown in boldface.
Table 7. RMSE and accuracy for various phases of test set T6. The best result for each metric is shown in boldface.
Method04590180
RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑)
DNNDOA [18]6.56 ± 15.5564.1027.78 ± 17.8646.2532.79 ± 17.8749.2922.77 ± 15.8242.36
MNOMP [63]13.76 ± 18.3953.3762.67 ± 20.1234.6758.91 ± 18.2932.0271.01 ± 18.0938.02
MTAPC [9]1.91 ± 7.3673.8722.97 ± 15.2956.5929.10 ± 12.2951.4218.91 ± 11.0158.12
MUSIC [5]1.32 ± 2.5691.2021.87 ± 15.2091.7124.09 ± 12.0992.0118.83 ± 12.5190.98
R-MUSIC [61]0.60 ± 1.3793.4527.38 ± 11.0591.7129.12 ± 11.4292.6134.10 ± 15.2189.08
Our0.27 ± 0.2199.587.12 ± 9.1292.1815.91 ± 14.6192.6119.06 ± 13.2490.49
Table 8. RMSE and accuracy for various frequencies. The best result for each metric is shown in boldface.
Table 8. RMSE and accuracy for various frequencies. The best result for each metric is shown in boldface.
Method1 GHz2.4 GHz3 GHz5 GHz10 GHz
RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑)
DNNDOA [18]6.56 ± 15.5564.1028.57 ± 16.8947.0127.39 ± 17.6945.2826.97 ± 18.0247.8627.20 ± 17.8345.35
MNOMP [63]13.76 ± 18.3953.3734.91 ± 21.2149.1234.20 ± 18.9151.2132.00 ± 23.1148.0331.93 ± 21.4339.55
MTAPC [9]1.91 ± 7.3673.8721.12 ± 15.2252.8619.78 ± 13.1752.5720.55 ± 13.0052.0020.44 ± 13.4749.88
MUSIC [5]1.32 ± 2.5691.2022.76 ± 14.6492.8823.11 ± 16.3791.5624.52 ± 16.4590.6324.80 ± 17.1391.38
R-MUSIC [61]0.60 ± 1.3793.4531.43 ± 11.5692.2230.92 ± 9.7691.5631.81 ± 9.6490.6332.53 ± 9.6391.52
Our0.27 ± 0.2199.5824.52 ± 13.7892.4521.46 ± 12.8093.0119.91 ± 12.4694.6318.56 ± 10.8292.74
Table 9. RMSE and accuracy for three modulation types in test set T8. The best result for each metric is shown in boldface.
Table 9. RMSE and accuracy for three modulation types in test set T8. The best result for each metric is shown in boldface.
MethodAMFMPSK
RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑) RMSE (↓) Acc. (↑)
DNNDOA [18]22.22 ± 12.8348.3826.02 ± 17.8245.4149.11 ± 21.8243.34
MNOMP [63]36.82 ± 23.6044.1231.69 ± 29.1447.2351.11 ± 11.2949.92
MTAPC [9]19.43 ± 8.4862.8822.91 ± 10.1749.8929.08 ± 15.2253.24
MUSIC [5]18.42 ± 11.1591.3722.84 ± 13.0293.9229.83 ± 17.1291.38
R-MUSIC [61]26.53 ± 9.6391.3727.91 ± 11.0192.8232.53 ± 9.6391.38
Our16.57 ± 11.7692.7419.82 ± 10.0592.2223.57 ± 12.9791.43
Table 10. Ablation results. The best result for each metric is shown in boldface.
Table 10. Ablation results. The best result for each metric is shown in boldface.
Model VersionRMSEAccuracy
Two-step training32.99 ± 9.6198.86
Loss function with λ 1 = 0.2 and λ 2 = 1.0 0.75 ± 0.4098.64
Loss function with λ 1 = 1.0 and λ 2 = 0.4 0.68 ± 0.3899.13
Our0.27 ± 0.2199.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bianco, S.; Celona, L.; Crotti, P.; Napoletano, P.; Petraglia, G.; Vinetti, P. Enhancing Direction-of-Arrival Estimation with Multi-Task Learning. Sensors 2024, 24, 7390. https://doi.org/10.3390/s24227390

AMA Style

Bianco S, Celona L, Crotti P, Napoletano P, Petraglia G, Vinetti P. Enhancing Direction-of-Arrival Estimation with Multi-Task Learning. Sensors. 2024; 24(22):7390. https://doi.org/10.3390/s24227390

Chicago/Turabian Style

Bianco, Simone, Luigi Celona, Paolo Crotti, Paolo Napoletano, Giovanni Petraglia, and Pietro Vinetti. 2024. "Enhancing Direction-of-Arrival Estimation with Multi-Task Learning" Sensors 24, no. 22: 7390. https://doi.org/10.3390/s24227390

APA Style

Bianco, S., Celona, L., Crotti, P., Napoletano, P., Petraglia, G., & Vinetti, P. (2024). Enhancing Direction-of-Arrival Estimation with Multi-Task Learning. Sensors, 24(22), 7390. https://doi.org/10.3390/s24227390

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop