764 research outputs found
Sample variance in photometric redshift calibration: cosmological biases and survey requirements
We use N-body/photometric galaxy simulations to examine the impact of sample variance of spectroscopic redshift samples on the accuracy of photometric redshift (photo-z) determination and calibration of photo-z errors. We estimate the biases in the cosmological parameter constraints from weak lensing and derive requirements on the spectroscopic follow-up for three different photo-z algorithms chosen to broadly span the range of algorithms available. We find that sample variance is much more relevant for the photo-z error calibration than for photo-z training, implying that follow-up requirements are similar for different algorithms. We demonstrate that the spectroscopic sample can be used for training of photo-zs and error calibration without incurring additional bias in the cosmological parameters. We provide a guide for observing proposals for the spectroscopic follow-up to ensure that redshift calibration biases do not dominate the cosmological parameter error budget. For example, assuming optimistically (pessimistically) that the weak lensing shear measurements from the Dark Energy Survey could obtain 1Ï constraints on the dark energy equation of state w of 0.035 (0.055), implies a follow-up requirement of 150 (40) patches of sky with a telescope such as Magellan, assuming a 1/8 deg2 effective field of view and 400 galaxies per patch. Assuming (optimistically) a VIMOS-VLT Deep Survey-like spectroscopic completeness with purely random failures, this could be accomplished with about 75 (20) nights of observation. For more realistic assumptions regarding spectroscopic completeness, or with the presence of other sources of systematics not considered here, further degradations to dark energy constraints are possible. We test several approaches for making the requirements less stringent. For example, if the redshift distribution of the overall sample can be estimated by some other technique, e.g. cross-correlation, then follow-up requirements could be reduced by an order of magnitud
A High Throughput Workflow Environment for Cosmological Simulations
The next generation of wide-area sky surveys offer the power to place
extremely precise constraints on cosmological parameters and to test the source
of cosmic acceleration. These observational programs will employ multiple
techniques based on a variety of statistical signatures of galaxies and
large-scale structure. These techniques have sources of systematic error that
need to be understood at the percent-level in order to fully leverage the power
of next-generation catalogs. Simulations of large-scale structure provide the
means to characterize these uncertainties. We are using XSEDE resources to
produce multiple synthetic sky surveys of galaxies and large-scale structure in
support of science analysis for the Dark Energy Survey. In order to scale up
our production to the level of fifty 10^10-particle simulations, we are working
to embed production control within the Apache Airavata workflow environment. We
explain our methods and report how the workflow has reduced production time by
40% compared to manual management.Comment: 8 pages, 5 figures. V2 corrects an error in figure
Cross-correlation Weak Lensing of SDSS Galaxy Clusters III: Mass-to-light Ratios
We present measurements of the excess mass-to-light ratio measured
aroundMaxBCG galaxy clusters observed in the SDSS. This red sequence cluster
sample includes objects from small groups with masses ranging from ~5x10^{12}
to ~10^{15} M_{sun}/h. Using cross-correlation weak lensing, we measure the
excess mass density profile above the universal mean \Delta \rho(r) = \rho(r) -
\bar{\rho} for clusters in bins of richness and optical luminosity. We also
measure the excess luminosity density \Delta l(r) = l(r) - \bar{l} measured in
the z=0.25 i-band. For both mass and light, we de-project the profiles to
produce 3D mass and light profiles over scales from 25 kpc/ to 22 Mpc/h. From
these profiles we calculate the cumulative excess mass M(r) and excess light
L(r) as a function of separation from the BCG. On small scales, where \rho(r)
>> \bar{\rho}, the integrated mass-to-light profile may be interpreted as the
cluster mass-to-light ratio. We find the M/L_{200}, the mass-to-light ratio
within r_{200}, scales with cluster mass as a power law with index 0.33+/-0.02.
On large scales, where \rho(r) ~ \bar{\rho}, the M/L approaches an asymptotic
value independent of cluster richness. For small groups, the mean M/L_{200} is
much smaller than the asymptotic value, while for large clusters it is
consistent with the asymptotic value. This asymptotic value should be
proportional to the mean mass-to-light ratio of the universe . We find
/b^2_{ml} = 362+/-54 h (statistical). There is additional uncertainty in
the overall calibration at the ~10% level. The parameter b_{ml} is primarily a
function of the bias of the L <~ L_* galaxies used as light tracers, and should
be of order unity. Multiplying by the luminosity density in the same bandpass
we find \Omega_m/b^2_{ml} = 0.02+/-0.03, independent of the Hubble parameter.Comment: Third paper in a series; v2.0 incorporates ApJ referee's suggestion
Future Evolution of Structure in an Accelerating Universe
Current cosmological data indicate that our universe contains a substantial
component of dark vacuum energy that is driving the cosmos to accelerate. We
examine the immediate and longer term consequences of this dark energy (assumed
here to have a constant density). Using analytic calculations and supporting
numerical simulations, we present criteria for test bodies to remain bound to
existing structures. We show that collapsed halos become spatially isolated and
dynamically relax to a particular density profile with logarithmic slope
steeper than -3 at radii beyond r_200. The asymptotic form of the space-time
metric is then specified. We develop this scenario further by determining the
effects of the accelerating expansion on the background radiation fields and
individual particles. In an appendix, we generalize these results to include
quintessence.Comment: 12 pages, 6 figures. Submitted to ApJ on April 23, 200
Cognitive reserve proxies do not differentially account for cognitive performance in patients with focal frontal and non-frontal lesions
Objective: Cognitive reserve (CR) suggests that premorbid efficacy, aptitude, and flexibility of cognitive processing can aid the brain\u2019s ability to cope with change or damage. Our previous work has shown that age and literacy attainment predict the cognitive performance of frontal patients on frontal-executive tests. However, it remains unknown whether CR also predicts the cognitive performance of non-frontal patients. Method: We investigated the independent effect of a CR proxy, National Adult Reading Test (NART) IQ, as well as age and lesion group (frontal vs. non-frontal) on measures of executive function, intelligence, processing speed, and naming in 166 patients with focal, unilateral frontal lesions; 91 patients with focal, unilateral non-frontal lesions; and 136 healthy controls. Results: Fitting multiple linear regression models for each cognitive measure revealed that NART IQ predicted executive, intelligence, and naming performance. Age also significantly predicted performance on the executive and processing speed tests. Finally, belonging to the frontal group predicted executive and naming performance, while membership of the non-frontal group predicted intelligence. Conclusions: These findings suggest that age, lesion group, and literacy attainment play independent roles in predicting cognitive performance following stroke or brain tumour. However, the relationship between CR and focal brain damage does not differ in the context of frontal and non-frontal lesions
The Dynamical State and Mass-Concentration Relation of Galaxy Clusters
We use the Millennium Simulation series to study how the dynamical state of
dark matter halos affects the relation between mass and concentration. We find
that a large fraction of massive systems are identified when they are
substantially out of equilibrium and in a particular phase of their dynamical
evolution: the more massive the halo, the more likely it is found at a
transient stage of high concentration. This state reflects the recent assembly
of massive halos and corresponds to the first pericentric passage of
recently-accreted material when, before virialization, the kinetic and
potential energies reach maximum and minimum values, respectively. This result
explains the puzzling upturn in the mass-concentration relation reported in
recent work for massive halos; indeed, the upturn disappears when only
dynamically-relaxed systems are considered in the analysis. Our results warn
against applying simple equilibrium models to describe the structure of rare,
massive galaxy clusters and urges caution when extrapolating scaling laws
calibrated on lower-mass systems, where such deviations from equilibrium are
less common. The evolving dynamical state of galaxy clusters ought to be
carefully taken into account if cluster studies are to provide precise
cosmological constraints.Comment: 8 Pages. Minor changes to match published versio
The asymptotic structure of space-time
Astronomical observations strongly suggest that our universe is now accelerating and contains a substantial admixture of dark vacuum energy. Using numerical simulations to study this newly consolidated cosmological model (with a constant density of dark energy), we show that astronomical structures freeze out in the near future and that the density profiles of dark matter halos approach the same general form. Every dark matter halo grows asymptotically isolated and thereby becomes the center of its own island universe. Each of these isolated regions of space-time approaches a universal geometry and we calculate the corresponding form of the space-time metric.Peer Reviewedhttp://deepblue.lib.umich.edu/bitstream/2027.42/60612/1/Adams2003Asymptotic.pd
Higher-order factors of personality: Do they exist?
Scales that measure the Big Five personality factors are often substantially intercorrelated. These correlations are sometimes interpreted as implying the existence of two higher order factors of personality. The authors show that correlations between measures of broad personality factors do not necessarily imply the existence of higher order factors and might instead be due to variables that represent same-signed blends of orthogonal factors. Therefore, the hypotheses of higher order factors and blended variables can only be tested with data on lower level personality variables that define the personality factors. The authors compared the higher order factor model and the blended variable model in three participant samples using the Big Five Aspect Scales, and found better fit for the latter model. In other analyses using the HEXACO Personality Inventory, they identified mutually uncorrelated markers of six personality factors. The authors conclude that correlations between personality factor scales can be explained without postulating any higher order dimensions of personality. © 2009 by the Society for Personality and Social Psychology, Inc
- âŠ