4,987 research outputs found
CMU DeepLens: Deep Learning For Automatic Image-based Galaxy-Galaxy Strong Lens Finding
Galaxy-scale strong gravitational lensing is not only a valuable probe of the
dark matter distribution of massive galaxies, but can also provide valuable
cosmological constraints, either by studying the population of strong lenses or
by measuring time delays in lensed quasars. Due to the rarity of galaxy-scale
strongly lensed systems, fast and reliable automated lens finding methods will
be essential in the era of large surveys such as LSST, Euclid, and WFIRST. To
tackle this challenge, we introduce CMU DeepLens, a new fully automated
galaxy-galaxy lens finding method based on Deep Learning. This supervised
machine learning approach does not require any tuning after the training step
which only requires realistic image simulations of strongly lensed systems. We
train and validate our model on a set of 20,000 LSST-like mock observations
including a range of lensed systems of various sizes and signal-to-noise ratios
(S/N). We find on our simulated data set that for a rejection rate of
non-lenses of 99%, a completeness of 90% can be achieved for lenses with
Einstein radii larger than 1.4" and S/N larger than 20 on individual -band
LSST exposures. Finally, we emphasize the importance of realistically complex
simulations for training such machine learning methods by demonstrating that
the performance of models of significantly different complexities cannot be
distinguished on simpler simulations. We make our code publicly available at
https://github.com/McWilliamsCenter/CMUDeepLens .Comment: 12 pages, 9 figures, submitted to MNRA
Personnel recognition and gait classification based on multistatic micro-doppler signatures using deep convolutional neural networks
In this letter, we propose two methods for personnel recognition and gait classification using deep convolutional neural networks (DCNNs) based on multistatic radar micro-Doppler signatures. Previous DCNN-based schemes have mainly focused on monostatic scenarios, whereas directional diversity offered by multistatic radar is exploited in this letter to improve classification accuracy. We first propose the voted monostatic DCNN (VMo-DCNN) method, which trains DCNNs on each receiver node separately and fuses the results by binary voting. By merging the fusion step into the network architecture, we further propose the multistatic DCNN (Mul-DCNN) method, which performs slightly better than VMo-DCNN. These methods are validated on real data measured with a 2.4-GHz multistatic radar system. Experimental results show that the Mul-DCNN achieves over 99% accuracy in armed/unarmed gait classification using only 20% training data and similar performance in two-class personnel recognition using 50% training data, which are higher than the accuracy obtained by performing DCNN on a single radar node
- …