109 research outputs found
Machine learning cosmological structure formation
We train a machine learning algorithm to learn cosmological structure
formation from N-body simulations. The algorithm infers the relationship
between the initial conditions and the final dark matter haloes, without the
need to introduce approximate halo collapse models. We gain insights into the
physics driving halo formation by evaluating the predictive performance of the
algorithm when provided with different types of information about the local
environment around dark matter particles. The algorithm learns to predict
whether or not dark matter particles will end up in haloes of a given mass
range, based on spherical overdensities. We show that the resulting predictions
match those of spherical collapse approximations such as extended
Press-Schechter theory. Additional information on the shape of the local
gravitational potential is not able to improve halo collapse predictions; the
linear density field contains sufficient information for the algorithm to also
reproduce ellipsoidal collapse predictions based on the Sheth-Tormen model. We
investigate the algorithm's performance in terms of halo mass and radial
position and perform blind analyses on independent initial conditions
realisations to demonstrate the generality of our results.Comment: 10 pages, 7 figures. Minor changes to match version published in
MNRAS. Accepted on 22/06/201
Extending BEAMS to incorporate correlated systematic uncertainties
New supernova surveys such as the Dark Energy Survey, Pan-STARRS and the LSST
will produce an unprecedented number of photometric supernova candidates, most
with no spectroscopic data. Avoiding biases in cosmological parameters due to
the resulting inevitable contamination from non-Ia supernovae can be achieved
with the BEAMS formalism, allowing for fully photometric supernova cosmology
studies. Here we extend BEAMS to deal with the case in which the supernovae are
correlated by systematic uncertainties. The analytical form of the full BEAMS
posterior requires evaluating 2^N terms, where N is the number of supernova
candidates. This `exponential catastrophe' is computationally unfeasible even
for N of order 100. We circumvent the exponential catastrophe by marginalising
numerically instead of analytically over the possible supernova types: we
augment the cosmological parameters with nuisance parameters describing the
covariance matrix and the types of all the supernovae, \tau_i, that we include
in our MCMC analysis. We show that this method deals well even with large,
unknown systematic uncertainties without a major increase in computational
time, whereas ignoring the correlations can lead to significant biases and
incorrect credible contours. We then compare the numerical marginalisation
technique with a perturbative expansion of the posterior based on the insight
that future surveys will have exquisite light curves and hence the probability
that a given candidate is a Type Ia will be close to unity or zero, for most
objects. Although this perturbative approach changes computation of the
posterior from a 2^N problem into an N^2 or N^3 one, we show that it leads to
biases in general through a small number of misclassifications, implying that
numerical marginalisation is superior.Comment: Resubmitted under married name Lochner (formally Knights). Version 3:
major changes, including a large scale analysis with thousands of MCMC
chains. Matches version published in JCAP. 23 pages, 8 figure
Towards the Future of Supernova Cosmology
For future surveys, spectroscopic follow-up for all supernovae will be
extremely difficult. However, one can use light curve fitters, to obtain the
probability that an object is a Type Ia. One may consider applying a
probability cut to the data, but we show that the resulting non-Ia
contamination can lead to biases in the estimation of cosmological parameters.
A different method, which allows the use of the full dataset and results in
unbiased cosmological parameter estimation, is Bayesian Estimation Applied to
Multiple Species (BEAMS). BEAMS is a Bayesian approach to the problem which
includes the uncertainty in the types in the evaluation of the posterior. Here
we outline the theory of BEAMS and demonstrate its effectiveness using both
simulated datasets and SDSS-II data. We also show that it is possible to use
BEAMS if the data are correlated, by introducing a numerical marginalisation
over the types of the objects. This is largely a pedagogical introduction to
BEAMS with references to the main BEAMS papers.Comment: Replaced under married name Lochner (formally Knights). 3 pages, 2
figures. To appear in the Proceedings of 13th Marcel Grossmann Meeting
(MG13), Stockholm, Sweden, 1-7 July 201
Enabling Unsupervised Discovery in Astronomical Images through Self-Supervised Representations
Unsupervised learning, a branch of machine learning that can operate on
unlabelled data, has proven to be a powerful tool for data exploration and
discovery in astronomy. As large surveys and new telescopes drive a rapid
increase in data size and richness, these techniques offer the promise of
discovering new classes of objects and of efficient sorting of data into
similar types. However, unsupervised learning techniques generally require
feature extraction to derive simple but informative representations of images.
In this paper, we explore the use of self-supervised deep learning as a method
of automated representation learning. We apply the algorithm Bootstrap Your Own
Latent (BYOL) to Galaxy Zoo DECaLS images to obtain a lower dimensional
representation of each galaxy, known as features. We briefly validate these
features using a small supervised classification problem. We then move on to
apply an automated clustering algorithm, demonstrating that this fully
unsupervised approach is able to successfully group together galaxies with
similar morphology. The same features prove useful for anomaly detection, where
we use the framework astronomaly to search for merger candidates. While the
focus of this work is on optical images, we also explore the versatility of
this technique by applying the exact same approach to a small radio galaxy
dataset. This work aims to demonstrate that applying deep representation
learning is key to unlocking the potential of unsupervised discovery in future
datasets from telescopes such as the Vera C. Rubin Observatory and the Square
Kilometre Array.Comment: 22 pages, 22 figures, comments welcom
Bayesian inference for radio observations
New telescopes like the Square Kilometre Array (SKA) will push into a new sensitivity regime and expose systematics, such as direction-dependent effects, that could previously be ignored. Current methods for handling such systematics rely on alternating best estimates of instrumental calibration and models of the underlying sky, which can lead to inadequate uncertainty estimates and biased results because any correlations between parameters are ignored. These deconvolution algorithms produce a single image that is assumed to be a true representation of the sky, when in fact it is just one realization of an infinite ensemble of images compatible with the noise in the data. In contrast, here we report a Bayesian formalism that simultaneously infers both systematics and science. Our technique, Bayesian Inference for Radio Observations (BIRO), determines all parameters directly from the raw data, bypassing image-making entirely, by sampling from the joint posterior probability distribution. This enables it to derive both correlations and accurate uncertainties, making use of the flexible software meqtrees to model the sky and telescope simultaneously. We demonstrate BIRO with two simulated sets of Westerbork Synthesis Radio Telescope data sets. In the first, we perform joint estimates of 103 scientific (flux densities of sources) and instrumental (pointing errors, beamwidth and noise) parameters. In the second example, we perform source separation with BIRO. Using the Bayesian evidence, we can accurately select between a single point source, two point sources and an extended Gaussian source, allowing for ‘super-resolution' on scales much smaller than the synthesized bea
Deep Multi-object Spectroscopy to Enhance Dark Energy Science from LSST
Community access to deep (i ~ 25), highly-multiplexed optical and
near-infrared multi-object spectroscopy (MOS) on 8-40m telescopes would greatly
improve measurements of cosmological parameters from LSST. The largest gain
would come from improvements to LSST photometric redshifts, which are employed
directly or indirectly for every major LSST cosmological probe; deep
spectroscopic datasets will enable reduced uncertainties in the redshifts of
individual objects via optimized training. Such spectroscopy will also
determine the relationship of galaxy SEDs to their environments, key
observables for studies of galaxy evolution. The resulting data will also
constrain the impact of blending on photo-z's. Focused spectroscopic campaigns
can also improve weak lensing cosmology by constraining the intrinsic
alignments between the orientations of galaxies. Galaxy cluster studies can be
enhanced by measuring motions of galaxies in and around clusters and by testing
photo-z performance in regions of high density. Photometric redshift and
intrinsic alignment studies are best-suited to instruments on large-aperture
telescopes with wider fields of view (e.g., Subaru/PFS, MSE, or GMT/MANIFEST)
but cluster investigations can be pursued with smaller-field instruments (e.g.,
Gemini/GMOS, Keck/DEIMOS, or TMT/WFOS), so deep MOS work can be distributed
amongst a variety of telescopes. However, community access to large amounts of
nights for surveys will still be needed to accomplish this work. In two
companion white papers we present gains from shallower, wide-area MOS and from
single-target imaging and spectroscopy.Comment: Science white paper submitted to the Astro2020 decadal survey. A
table of time requirements is available at
http://d-scholarship.pitt.edu/36036
- …
