115 research outputs found
Recommended from our members
Fusing Deep Learning and Sparse Coding for SAR ATR
We propose a multimodal and multidiscipline data fusion strategy appropriate for automatic target recognition (ATR) on synthetic aperture radar imagery. Our architecture fuses a proposed clustered version of the AlexNet convolutional neural network with sparse coding theory that is extended to facilitate an adaptive elastic net optimization concept. Evaluation on the MSTAR dataset yields the highest ATR performance reported yet, which is 99.33% and 99.86% for the three- and ten-class problems, respectively
Target recognition for synthetic aperture radar imagery based on convolutional neural network feature fusion
Driven by the great success of deep convolutional neural networks (CNNs) that are currently used by quite a few computer vision applications, we extend the usability of visual-based CNNs into the synthetic aperture radar (SAR) data domain without employing transfer learning. Our SAR automatic target recognition (ATR) architecture efficiently extends the pretrained Visual Geometry Group CNN from the visual domain into the X-band SAR data domain by clustering its neuron layers, bridging the visual—SAR modality gap by fusing the features extracted from the hidden layers, and by employing a local feature matching scheme. Trials on the moving and stationary target acquisition dataset under various setups and nuisances demonstrate a highly appealing ATR performance gaining 100% and 99.79% in the 3-class and 10-class ATR problem, respectively. We also confirm the validity, robustness, and conceptual coherence of the proposed method by extending it to several state-of-the-art CNNs and commonly used local feature similarity/match metrics
Sparse Signal Models for Data Augmentation in Deep Learning ATR
Automatic Target Recognition (ATR) algorithms classify a given Synthetic
Aperture Radar (SAR) image into one of the known target classes using a set of
training images available for each class. Recently, learning methods have shown
to achieve state-of-the-art classification accuracy if abundant training data
is available, sampled uniformly over the classes, and their poses. In this
paper, we consider the task of ATR with a limited set of training images. We
propose a data augmentation approach to incorporate domain knowledge and
improve the generalization power of a data-intensive learning algorithm, such
as a Convolutional neural network (CNN). The proposed data augmentation method
employs a limited persistence sparse modeling approach, capitalizing on
commonly observed characteristics of wide-angle synthetic aperture radar (SAR)
imagery. Specifically, we exploit the sparsity of the scattering centers in the
spatial domain and the smoothly-varying structure of the scattering
coefficients in the azimuthal domain to solve the ill-posed problem of
over-parametrized model fitting. Using this estimated model, we synthesize new
images at poses and sub-pixel translations not available in the given data to
augment CNN's training data. The experimental results show that for the
training data starved region, the proposed method provides a significant gain
in the resulting ATR algorithm's generalization performance.Comment: 12 pages, 5 figures, to be submitted to IEEE Transactions on
Geoscience and Remote Sensin
Hierarchical Disentanglement-Alignment Network for Robust SAR Vehicle Recognition
Vehicle recognition is a fundamental problem in SAR image interpretation.
However, robustly recognizing vehicle targets is a challenging task in SAR due
to the large intraclass variations and small interclass variations.
Additionally, the lack of large datasets further complicates the task. Inspired
by the analysis of target signature variations and deep learning
explainability, this paper proposes a novel domain alignment framework named
the Hierarchical Disentanglement-Alignment Network (HDANet) to achieve
robustness under various operating conditions. Concisely, HDANet integrates
feature disentanglement and alignment into a unified framework with three
modules: domain data generation, multitask-assisted mask disentanglement, and
domain alignment of target features. The first module generates diverse data
for alignment, and three simple but effective data augmentation methods are
designed to simulate target signature variations. The second module
disentangles the target features from background clutter using the
multitask-assisted mask to prevent clutter from interfering with subsequent
alignment. The third module employs a contrastive loss for domain alignment to
extract robust target features from generated diverse data and disentangled
features. Lastly, the proposed method demonstrates impressive robustness across
nine operating conditions in the MSTAR dataset, and extensive qualitative and
quantitative analyses validate the effectiveness of our framework
EMC2A-Net: An Efficient Multibranch Cross-channel Attention Network for SAR Target Classification
In recent years, convolutional neural networks (CNNs) have shown great
potential in synthetic aperture radar (SAR) target recognition. SAR images have
a strong sense of granularity and have different scales of texture features,
such as speckle noise, target dominant scatterers and target contours, which
are rarely considered in the traditional CNN model. This paper proposed two
residual blocks, namely EMC2A blocks with multiscale receptive fields(RFs),
based on a multibranch structure and then designed an efficient isotopic
architecture deep CNN (DCNN), EMC2A-Net. EMC2A blocks utilize parallel dilated
convolution with different dilation rates, which can effectively capture
multiscale context features without significantly increasing the computational
burden. To further improve the efficiency of multiscale feature fusion, this
paper proposed a multiscale feature cross-channel attention module, namely the
EMC2A module, adopting a local multiscale feature interaction strategy without
dimensionality reduction. This strategy adaptively adjusts the weights of each
channel through efficient one-dimensional (1D)-circular convolution and sigmoid
function to guide attention at the global channel wise level. The comparative
results on the MSTAR dataset show that EMC2A-Net outperforms the existing
available models of the same type and has relatively lightweight network
structure. The ablation experiment results show that the EMC2A module
significantly improves the performance of the model by using only a few
parameters and appropriate cross-channel interactions.Comment: 15 pages, 9 figures, Submitted to IEEE Transactions on Geoscience and
Remote Sensing, 202
A three-step classification framework to handle complex data distribution for radar UAV detection
Unmanned aerial vehicles (UAVs) have been used in a wide range of applications and become an increasingly important radar target. To better model radar data and to tackle the curse of dimensionality, a three-step classification framework is proposed for UAV detection. First we propose to utilize the greedy subspace clustering to handle potential outliers and the complex sample distribution of radar data. Parameters of the resulting multi-Gaussian model, especially the covariance matrices, could not be reliably estimated due to insufficient training samples and the high dimensionality. Thus, in the second step, a multi-Gaussian subspace reliability analysis is proposed to handle the unreliable feature dimensions of these covariance matrices. To address the challenges of classifying samples using the complex multi-Gaussian model and to fuse the distances of a sample to different clusters at different dimensionalities, a subspace-fusion scheme is proposed in the third step. The proposed approach is validated on a large benchmark dataset, which significantly outperforms the state-of-the-art approaches
- …