12,089 research outputs found

    Search strategies for long gravitational-wave transients: hidden Markov model tracking and seedless clustering

    Get PDF
    A number of detections have been made in the past few years of gravitational waves from compact binary coalescences. While there exist well-understood waveform models for signals from compact binary coalescences, many sources of gravitational waves are not well modeled, including potential long-transient signals from a binary neutron star post-merger remnant. Searching for these sources requires robust detection algorithms that make minimal assumptions about any potential signals. In this paper, we compare two unmodeled search schemes for long-transient gravitational waves, operating on cross-power spectrograms. One is an efficient algorithm first implemented for continuous wave searches, based on a hidden Markov model. The other is a seedless clustering method, which has been used in transient gravitational wave analysis in the past. We quantify the performance of both algorithms, including sensitivity and computational cost, by simulating synthetic signals with a special focus on sources like binary neutron star post-merger remnants. We demonstrate that the hidden Markov model tracking is a good option in model-agnostic searches for low signal-to-noise ratio signals. We also show that it can outperform the seedless method for certain categories of signals while also being computationally more efficient.Comment: 10 pages, 7 figure

    Practical recommendations for gradient-based training of deep architectures

    Full text link
    Learning algorithms related to artificial neural networks and in particular for Deep Learning may seem to involve many bells and whistles, called hyper-parameters. This chapter is meant as a practical guide with recommendations for some of the most commonly used hyper-parameters, in particular in the context of learning algorithms based on back-propagated gradient and gradient-based optimization. It also discusses how to deal with the fact that more interesting results can be obtained when allowing one to adjust many hyper-parameters. Overall, it describes elements of the practice used to successfully and efficiently train and debug large-scale and often deep multi-layer neural networks. It closes with open questions about the training difficulties observed with deeper architectures
    • …
    corecore