12,089 research outputs found
Search strategies for long gravitational-wave transients: hidden Markov model tracking and seedless clustering
A number of detections have been made in the past few years of gravitational
waves from compact binary coalescences. While there exist well-understood
waveform models for signals from compact binary coalescences, many sources of
gravitational waves are not well modeled, including potential long-transient
signals from a binary neutron star post-merger remnant. Searching for these
sources requires robust detection algorithms that make minimal assumptions
about any potential signals. In this paper, we compare two unmodeled search
schemes for long-transient gravitational waves, operating on cross-power
spectrograms. One is an efficient algorithm first implemented for continuous
wave searches, based on a hidden Markov model. The other is a seedless
clustering method, which has been used in transient gravitational wave analysis
in the past. We quantify the performance of both algorithms, including
sensitivity and computational cost, by simulating synthetic signals with a
special focus on sources like binary neutron star post-merger remnants. We
demonstrate that the hidden Markov model tracking is a good option in
model-agnostic searches for low signal-to-noise ratio signals. We also show
that it can outperform the seedless method for certain categories of signals
while also being computationally more efficient.Comment: 10 pages, 7 figure
Practical recommendations for gradient-based training of deep architectures
Learning algorithms related to artificial neural networks and in particular
for Deep Learning may seem to involve many bells and whistles, called
hyper-parameters. This chapter is meant as a practical guide with
recommendations for some of the most commonly used hyper-parameters, in
particular in the context of learning algorithms based on back-propagated
gradient and gradient-based optimization. It also discusses how to deal with
the fact that more interesting results can be obtained when allowing one to
adjust many hyper-parameters. Overall, it describes elements of the practice
used to successfully and efficiently train and debug large-scale and often deep
multi-layer neural networks. It closes with open questions about the training
difficulties observed with deeper architectures
- …