3,887 research outputs found
Data Provenance and Management in Radio Astronomy: A Stream Computing Approach
New approaches for data provenance and data management (DPDM) are required
for mega science projects like the Square Kilometer Array, characterized by
extremely large data volume and intense data rates, therefore demanding
innovative and highly efficient computational paradigms. In this context, we
explore a stream-computing approach with the emphasis on the use of
accelerators. In particular, we make use of a new generation of high
performance stream-based parallelization middleware known as InfoSphere
Streams. Its viability for managing and ensuring interoperability and integrity
of signal processing data pipelines is demonstrated in radio astronomy. IBM
InfoSphere Streams embraces the stream-computing paradigm. It is a shift from
conventional data mining techniques (involving analysis of existing data from
databases) towards real-time analytic processing. We discuss using InfoSphere
Streams for effective DPDM in radio astronomy and propose a way in which
InfoSphere Streams can be utilized for large antennae arrays. We present a
case-study: the InfoSphere Streams implementation of an autocorrelating
spectrometer, and using this example we discuss the advantages of the
stream-computing approach and the utilization of hardware accelerators
A GPU based real-time software correlation system for the Murchison Widefield Array prototype
Modern graphics processing units (GPUs) are inexpensive commodity hardware
that offer Tflop/s theoretical computing capacity. GPUs are well suited to many
compute-intensive tasks including digital signal processing.
We describe the implementation and performance of a GPU-based digital
correlator for radio astronomy. The correlator is implemented using the NVIDIA
CUDA development environment. We evaluate three design options on two
generations of NVIDIA hardware. The different designs utilize the internal
registers, shared memory and multiprocessors in different ways. We find that
optimal performance is achieved with the design that minimizes global memory
reads on recent generations of hardware.
The GPU-based correlator outperforms a single-threaded CPU equivalent by a
factor of 60 for a 32 antenna array, and runs on commodity PC hardware. The
extra compute capability provided by the GPU maximises the correlation
capability of a PC while retaining the fast development time associated with
using standard hardware, networking and programming languages. In this way, a
GPU-based correlation system represents a middle ground in design space between
high performance, custom built hardware and pure CPU-based software
correlation.
The correlator was deployed at the Murchison Widefield Array 32 antenna
prototype system where it ran in real-time for extended periods. We briefly
describe the data capture, streaming and correlation system for the prototype
array.Comment: 11 pages, to appear in PAS
Observations of transients and pulsars with LOFAR international stations
The LOw FRequency ARray - LOFAR is a new radio telescope that is moving the
science of radio pulsars and transients into a new phase. Its design places
emphasis on digital hardware and flexible software instead of mechanical
solutions. LOFAR observes at radio frequencies between 10 and 240 MHz where
radio pulsars and many transients are expected to be brightest. Radio frequency
signals emitted from these objects allow us to study the intrinsic pulsar
emission and phenomena such as propagation effects through the interstellar
medium. The design of LOFAR allows independent use of its stations to conduct
observations of known bright objects, or wide field monitoring of transient
events. One such combined software/hardware solution is called the Advanced
Radio Transient Event Monitor and Identification System (ARTEMIS). It is a
backend for both targeted observations and real-time searches for millisecond
radio transients which uses Graphical Processing Unit (GPU) technology to
remove interstellar dispersion and detect millisecond radio bursts from
astronomical sources in real-time using a single LOFAR station.Comment: To appear in the proceedings of the Electromagnetic Radiation from
Pulsars and Magnetars conference, Zielona Gora, 2012. 4 pages, 1 figur
The UTMOST: A hybrid digital signal processor transforms the MOST
The Molonglo Observatory Synthesis Telescope (MOST) is an 18,000 square meter
radio telescope situated some 40 km from the city of Canberra, Australia. Its
operating band (820-850 MHz) is now partly allocated to mobile phone
communications, making radio astronomy challenging. We describe how the
deployment of new digital receivers (RX boxes), Field Programmable Gate Array
(FPGA) based filterbanks and server-class computers equipped with 43 GPUs
(Graphics Processing Units) has transformed MOST into a versatile new
instrument (the UTMOST) for studying the dynamic radio sky on millisecond
timescales, ideal for work on pulsars and Fast Radio Bursts (FRBs). The
filterbanks, servers and their high-speed, low-latency network form part of a
hybrid solution to the observatory's signal processing requirements. The
emphasis on software and commodity off-the-shelf hardware has enabled rapid
deployment through the re-use of proven 'software backends' for its signal
processing. The new receivers have ten times the bandwidth of the original MOST
and double the sampling of the line feed, which doubles the field of view. The
UTMOST can simultaneously excise interference, make maps, coherently dedisperse
pulsars, and perform real-time searches of coherent fan beams for dispersed
single pulses. Although system performance is still sub-optimal, a pulsar
timing and FRB search programme has commenced and the first UTMOST maps have
been made. The telescope operates as a robotic facility, deciding how to
efficiently target pulsars and how long to stay on source, via feedback from
real-time pulsar folding. The regular timing of over 300 pulsars has resulted
in the discovery of 7 pulsar glitches and 3 FRBs. The UTMOST demonstrates that
if sufficient signal processing can be applied to the voltage streams it is
possible to perform innovative radio science in hostile radio frequency
environments.Comment: 12 pages, 6 figure
A Scalable Correlator Architecture Based on Modular FPGA Hardware, Reuseable Gateware, and Data Packetization
A new generation of radio telescopes is achieving unprecedented levels of
sensitivity and resolution, as well as increased agility and field-of-view, by
employing high-performance digital signal processing hardware to phase and
correlate large numbers of antennas. The computational demands of these imaging
systems scale in proportion to BMN^2, where B is the signal bandwidth, M is the
number of independent beams, and N is the number of antennas. The
specifications of many new arrays lead to demands in excess of tens of PetaOps
per second.
To meet this challenge, we have developed a general purpose correlator
architecture using standard 10-Gbit Ethernet switches to pass data between
flexible hardware modules containing Field Programmable Gate Array (FPGA)
chips. These chips are programmed using open-source signal processing libraries
we have developed to be flexible, scalable, and chip-independent. This work
reduces the time and cost of implementing a wide range of signal processing
systems, with correlators foremost among them,and facilitates upgrading to new
generations of processing technology. We present several correlator
deployments, including a 16-antenna, 200-MHz bandwidth, 4-bit, full Stokes
parameter application deployed on the Precision Array for Probing the Epoch of
Reionization.Comment: Accepted to Publications of the Astronomy Society of the Pacific. 31
pages. v2: corrected typo, v3: corrected Fig. 1
DiFX2: A more flexible, efficient, robust and powerful software correlator
Software correlation, where a correlation algorithm written in a high-level
language such as C++ is run on commodity computer hardware, has become
increasingly attractive for small to medium sized and/or bandwidth constrained
radio interferometers. In particular, many long baseline arrays (which
typically have fewer than 20 elements and are restricted in observing bandwidth
by costly recording hardware and media) have utilized software correlators for
rapid, cost-effective correlator upgrades to allow compatibility with new,
wider bandwidth recording systems and improve correlator flexibility. The DiFX
correlator, made publicly available in 2007, has been a popular choice in such
upgrades and is now used for production correlation by a number of
observatories and research groups worldwide. Here we describe the evolution in
the capabilities of the DiFX correlator over the past three years, including a
number of new capabilities, substantial performance improvements, and a large
amount of supporting infrastructure to ease use of the code. New capabilities
include the ability to correlate a large number of phase centers in a single
correlation pass, the extraction of phase calibration tones, correlation of
disparate but overlapping sub-bands, the production of rapidly sampled
filterbank and kurtosis data at minimal cost, and many more. The latest version
of the code is at least 15% faster than the original, and in certain situations
many times this value. Finally, we also present detailed test results
validating the correctness of the new code.Comment: 28 pages, 9 figures, accepted for publication in PAS
- …