78,272 research outputs found
Probing the first galaxies with the SKA
Observations of anisotropies in the brightness temperature of the 21 cm line
of neutral hydrogen from the period before reionization would shed light on the
dawn of the first stars and galaxies. In this paper, we use large-scale
semi-numerical simulations to analyse the imprint on the 21 cm signal of
spatial fluctuations in the Lyman-alpha flux arising from the clustering of the
first galaxies. We show that an experiment such as the Square Kilometer Array
(SKA) can probe this signal at the onset of reionization, giving us important
information about the UV emission spectra of the first stars and characterizing
their host galaxies. SKA-pathfinders with ~ 10% of the full collecting area
should be capable of making a statistical detection of the 21 cm power spectrum
at redshifts z 67 MHz). We then show
that the SKA should be able to measure the three dimensional power spectrum as
a function of the angle with the line of sight and discuss the use of the
redshift space distortions as a way to separate out the different components of
the 21 cm power spectrum. We demonstrate that, at least on large scales where
the Lyman-alpha fluctuations are linear, they can be used as a model
independent way to extract the power spectra due to these Lyman-alpha
fluctuations.Comment: 13 pages, 17 figures. New version to match version accepted by A&A.
Improved discussions on the Lyman-alpha simulation, adiabatic cooling
fluctuations, the Fisher matrix approach and the Poisson term calculation.
New version of the code available at: http://www.SimFast21.or
Forecasts for Dark Energy Measurements with Future HI Surveys
We use two independent methods to forecast the dark energy measurements
achievable by combining future galaxy redshift surveys based on the radio HI
emission line with Cosmic Microwave Background (CMB) data from the {\sl Planck}
satellite. In the first method we focus on the `standard ruler' provided by the
baryon acoustic oscillation (BAO) length scale. In the second method we utilize
additional information encoded in the galaxy power spectrum including galaxy
bias from velocity-space distortions and the growth of cosmic structure. We
find that a radio synthesis array with about 10 per cent of the collecting area
of the Square Kilometre Array (SKA), equipped with a wide () field-of-view, would have the capacity to perform a redshift survey to a maximum redshift and
thereby produce dark energy measurements that are competitive with surveys
likely to be undertaken by optical telescopes around 2015. There would then be
powerful arguments for adding collecting area to such a `Phase-1' SKA because
of the square-law scaling of survey speed with telescope sensitivity for HI
surveys, compared to the linear scaling for optical redshift surveys. The full
SKA telescope should, by performing a HI redshift
survey to around 2020, yield an accurate measurement of
cosmological parameters independent of CMB datasets. Combining CMB ({\sl
Planck}) and galaxy power spectrum (SKA) measurements will drive errors in the
dark energy equation-of-state parameter well below the 1 per cent level.
The major systematic uncertainty in these forecasts is the lack of direct
information about the mass function of high-redshift HI-emitting galaxies.Comment: 19 pages; 2 tables; 18 figures. accepted by MNRA
What Next-Generation 21 cm Power Spectrum Measurements Can Teach Us About the Epoch of Reionization
A number of experiments are currently working towards a measurement of the 21
cm signal from the Epoch of Reionization. Whether or not these experiments
deliver a detection of cosmological emission, their limited sensitivity will
prevent them from providing detailed information about the astrophysics of
reionization. In this work, we consider what types of measurements will be
enabled by a next-generation of larger 21 cm EoR telescopes. To calculate the
type of constraints that will be possible with such arrays, we use simple
models for the instrument, foreground emission, and the reionization history.
We focus primarily on an instrument modeled after the
collecting area Hydrogen Epoch of Reionization Array (HERA) concept design, and
parameterize the uncertainties with regard to foreground emission by
considering different limits to the recently described "wedge" footprint in
k-space. Uncertainties in the reionization history are accounted for using a
series of simulations which vary the ionizing efficiency and minimum virial
temperature of the galaxies responsible for reionization, as well as the mean
free path of ionizing photons through the IGM. Given various combinations of
models, we consider the significance of the possible power spectrum detections,
the ability to trace the power spectrum evolution versus redshift, the
detectability of salient power spectrum features, and the achievable level of
quantitative constraints on astrophysical parameters. Ultimately, we find that
of collecting area is enough to ensure a very high significance
() detection of the reionization power spectrum in even the
most pessimistic scenarios. This sensitivity should allow for meaningful
constraints on the reionization history and astrophysical parameters,
especially if foreground subtraction techniques can be improved and
successfully implemented.Comment: 27 pages, 18 figures, updated SKA numbers in appendi
A New Measurement of Cosmic Ray Composition at the Knee
The Dual Imaging Cerenkov Experiment (DICE) was designed and operated for
making elemental composition measurements of cosmic rays near the knee of the
spectrum at several PeV. Here we present the first results using this
experiment from the measurement of the average location of the depth of shower
maximum, , in the atmosphere as a function of particle energy. The value
of near the instrument threshold of ~0.1 PeV is consistent with
expectations from previous direct measurements. At higher energies there is
little change in composition up to ~5 PeV. Above this energy is deeper
than expected for a constant elemental composition implying the overall
elemental composition is becoming lighter above the knee region. These results
disagree with the idea that cosmic rays should become on average heavier above
the knee. Instead they suggest a transition to a qualitatively different
population of particles above 5 PeV.Comment: 7 pages, LaTeX, two eps figures, aas2pp4.sty and epsf.sty included,
accepted by Ap.J. Let
HI Epoch of Reionization Arrays
There are few data available with which to constrain the thermal history of
the intergalactic medium (IGM) following global recombination. Thus far, most
constraints flow from analyses of the Cosmic Microwave Background and optical
spectroscopy along a few lines of sight. However, direct study of the IGM in
emission or absorption against the CMB via the 1S hyperfine transition of
Hydrogen would enable broad characterization thermal history and source
populations. New generations of radio arrays are in development to measure this
line signature. Bright foreground emission and the complexity of instrument
calibration models are significant hurdles. How to optimize these is uncertain,
resulting in a diversity in approaches. We discuss recent limits on line
brightness, array efforts including the new Large Aperture Experiment to Detect
the Dark Ages (LEDA), and the next generation Hydrogen Reionization Array
(HERA) concept.Comment: 8 pages, 4 figures, 1 table. Invited review to the 11th Asian-Pacific
Regional IAU Meeting 2011, NARIT Conference Series, Vol. 1 eds. S.
Komonjinda, Y. Kovalev, and D. Ruffolo (2012
The First Release COSMOS Optical and Near-IR Data and Catalog
We present imaging data and photometry for the COSMOS survey in 15
photometric bands between 0.3um and 2.4um. These include data taken on the
Subaru 8.3m telescope, the KPNO and CTIO 4m telescopes, and the CFHT 3.6m
telescope. Special techniques are used to ensure that the relative photometric
calibration is better than 1% across the field of view. The absolute
photometric accuracy from standard star measurements is found to be 6%. The
absolute calibration is corrected using galaxy spectra, providing colors
accurate to 2% or better. Stellar and galaxy colors and counts agree well with
the expected values. Finally, as the first step in the scientific analysis of
these data we construct panchromatic number counts which confirm that both the
geometry of the universe and the galaxy population are evolving.Comment: 19 pages, 13 figures, 14 tables, Accepted to ApJS for COSMOS speciall
issu
The VLA Low-frequency Sky Survey
The Very Large Array (VLA) Low-frequency Sky Survey (VLSS) has imaged 95% of
the 3*pi sr of sky north of declination = -30 degrees at a frequency of 74 MHz
(4 meter wavelength). The resolution is 80" (FWHM) throughout, and the typical
RMS noise level is ~0.1 Jy/beam. The typical point-source detection limit is
0.7 Jy/beam and so far nearly 70,000 sources have been catalogued. This survey
used the 74 MHz system added to the VLA in 1998. It required new imaging
algorithms to remove the large ionospheric distortions at this very low
frequency throughout the entire ~11.9 degree field of view. This paper
describes the observation and data reduction methods used for the VLSS and
presents the survey images and source catalog. All of the calibrated images and
the source catalog are available online (http://lwa.nrl.navy.mil/VLSS) for use
by the astronomical community.Comment: 53 pages, including 3 tables and 15 figures. Has been accepted for
publication in the Astronomical Journa
A methodology for collecting valid software engineering data
An effective data collection method for evaluating software development methodologies and for studying the software development process is described. The method uses goal-directed data collection to evaluate methodologies with respect to the claims made for them. Such claims are used as a basis for defining the goals of the data collection, establishing a list of questions of interest to be answered by data analysis, defining a set of data categorization schemes, and designing a data collection form. The data to be collected are based on the changes made to the software during development, and are obtained when the changes are made. To insure accuracy of the data, validation is performed concurrently with software development and data collection. Validation is based on interviews with those people supplying the data. Results from using the methodology show that data validation is a necessary part of change data collection. Without it, as much as 50% of the data may be erroneous. Feasibility of the data collection methodology was demonstrated by applying it to five different projects in two different environments. The application showed that the methodology was both feasible and useful
- …