5,430 research outputs found
A new development cycle of the Statistical Toolkit
The Statistical Toolkit is an open source system specialized in the
statistical comparison of distributions. It addresses requirements common to
different experimental domains, such as simulation validation (e.g. comparison
of experimental and simulated distributions), regression testing in the course
of the software development process, and detector performance monitoring.
Various sets of statistical tests have been added to the existing collection to
deal with the one sample problem (i.e. the comparison of a data distribution to
a function, including tests for normality, categorical analysis and the
estimate of randomness). Improved algorithms and software design contribute to
the robustness of the results. A simple user layer dealing with primitive data
types facilitates the use of the toolkit both in standalone analyses and in
large scale experiments.Comment: To be published in the Proc. of CHEP (Computing in High Energy
Physics) 201
Bayesian reweighting of nuclear PDFs and constraints from proton-lead collisions at the LHC
New hard-scattering measurements from the LHC proton-lead run have the
potential to provide important constraints on the nuclear parton distributions
and thus contributing to a better understanding of the initial state in heavy
ion collisions. In order to quantify these constraints, as well as to assess
the compatibility with available nuclear data from fixed target experiments and
from RHIC, the traditional strategy is to perform a global fit of nuclear PDFs.
This procedure is however time consuming and technically challenging, and
moreover can only be performed by the PDF fitters themselves. In the case of
proton PDFs, an alternative approach has been suggested that uses Bayesian
inference to propagate the effects of new data into the PDFs without the need
of refitting. In this work, we apply this reweighting procedure to study the
impact on nuclear PDFs of low-mass Drell-Yan and single-inclusive
hadroproduction pseudo-data from proton-lead collisions at the LHC as
representative examples. In the hadroproduction case, in addition we assess the
possibility of discriminating between the DGLAP and CGC production frameworks.
We find that the LHC proton-lead data could lead to a substantial reduction of
the uncertainties on nuclear PDFs, in particular for the small-x gluon PDF
where uncertainties could decrease by up to a factor two. The Monte Carlo
replicas of EPS09 used in the analysis are released as a public code for
general use. It can be directly used, in particular, by the experimental
collaborations to check, in a straightforward manner, the degree of
compatibility of the new data with the global nPDF analyses.Comment: 21 pages, 10 figure
Distributional Dominance with Dirty Data
Distributional dominance criteria are commonly applied to draw welfare inferences about comparisons, but conclusions drawn from empirical implementations of dominance criteria may be influenced by data contamination. We examine a non-parametric approach to refining Lorenz-type comparisons and apply the technique to two important examples from the LIS data-base.Distributional dominance, Lorenz curve, robustness.
Robust Lorenz Curves: A Semiparametric Approach
Lorenz curves and second-order dominance criteria are known to be sensitive to data contamination in the right tail of the distribution. We propose two ways of dealing with the problem: (1) Estimate Lorenz curves using parametric models for income distributions, and (2) Combine empirical estimation with a parametric (robust) estimation of the upper tail of the distribution using the Pareto model. Approach (2) is preferred because of its flexibility. Using simulations we show the dramatic effect of a few contaminated data on the Lorenz ranking and the performance of the robust approach (2). Statistical inference tools are also provided.Welfare dominance, Lorenz curve, Pareto model, M-estimators.
Statistical Inference for Lorenz Curves with Censored Data
Lorenz curves and associated tools for ranking income distributions are commonly estimated on the assumption that full, unbiased samples are available. However, it is common to find income and wealth distributions that are routinely censored or trimmed. We derive the sampling distribution for a key family of statistics in the case where data have been modified in this fashion.Lorenz curve, sampling errors
Modelling Lorenz Curves:robust and semi-parametric issues
Modelling Lorenz curves (LC) for stochastic dominance comparisons is central to the analysis of income distribution. It is conventional to use non-parametric statistics based on empirical income cumulants which are in the construction of LC and other related second-order dominance criteria. However, although attractive because of its simplicity and its apparent flexibility, this approach suffers from important drawbacks. While no assumptions need to be made regarding the data-generating process (income distribution model), the empirical LC can be very sensitive to data particularities, especially in the upper tail of the distribution. This robustness problem can lead in practice to 'wrong' interpretation of dominance orders. A possible remedy for this problem is the use of parametric or semi-parametric models for the datagenerating process and robust estimators to obtain parameter estimates. In this paper, we focus on the robust estimation of semi parametric LC and investigate issues such as sensitivity of LC estimators to data contamination (Cowell and Victoria-Feser 2002), trimmed LC (Cowell and Victoria-Feser 2006) and inference for trimmed LC (Cowell and Victoria-Feser 2003), robust semi-parametric estimation for LC (Cowell and Victoria-Feser 2007) selection of optimal thresholds for (robust) semi parametric modelling (Dupuis and Victoria-Feser 2006) and use both simulations and real data to illustrate these points.
Interpreting Spectral Energy Distributions from Young Stellar Objects. I. A grid of 200,000 YSO model SEDs
We present a grid of radiation transfer models of axisymmetric young stellar
objects (YSOs), covering a wide range of stellar masses (from 0.1Msun to
50Msun) and evolutionary stages (from the early envelope infall stage to the
late disk-only stage). The grid consists of 20,000 YSO models, with spectral
energy distributions (SEDs) and polarization spectra computed at ten viewing
angles for each model, resulting in a total of 200,000 SEDs. [...]. These
models are publicly available on a dedicated WWW server:
http://www.astro.wisc.edu/protostars/ . In this paper we summarize the main
features of our models, as well as the range of parameters explored. [...]. We
examine the dependence of the spectral indices of the model SEDs on envelope
accretion rate and disk mass. In addition, we show variations of spectral
indices with stellar temperature, disk inner radius, and disk flaring power for
a subset of disk-only models. We also examine how changing the wavelength range
of data used to calculate spectral indices affects their values. We show sample
color-color plots of the entire grid as well as simulated clusters at various
distances with typical {\it Spitzer Space Telescope} sensitivities. We find
that young embedded sources generally occupy a large region of color-color
space due to inclination and stellar temperature effects. Disk sources occupy a
smaller region of color-color space, but overlap substantially with the region
occupied by embedded sources, especially in the near- and mid-IR. We identify
regions in color-color space where our models indicate that only sources at a
given evolutionary stage should lie. [...].Comment: 69 pages, 28 figures, Accepted for publication in ApJS. Preprint with
full resolution figures available at http://www.astro.wisc.edu/protostars
Validating delta-filters for resonant bar detectors of improved bandwidth foreseeing the future coincidence with interferometers
The classical delta filters used in the current resonant bar experiments for
detecting GW bursts are viable when the bandwidth of resonant bars is few Hz.
In that case, the incoming GW burst is likely to be viewed as an impulsive
signal in a very narrow frequency window. After making improvements in the
read-out with new transducers and high sensitivity dc-SQUID, the
Explorer-Nautilus have improved the bandwidth ( Hz) at the sensitivity
level of . Thus, it is necessary to reassess this
assumption of delta-like signals while building filters in the resonant bars as
the filtered output crucially depends on the shape of the waveform. This is
presented with an example of GW signals -- stellar quasi-normal modes, by
estimating the loss in SNR and the error in the timing, when the GW signal is
filtered with the delta filter as compared to the optimal filter.Comment: 7 pages, presented in Amaldi6, accepted for publication in Journal of
Physics: Conference Serie
A Comparative Study of Dark Energy Constraints from Current Observational Data
We examine how dark energy constraints from current observational data depend
on the analysis methods used: the analysis of Type Ia supernovae (SNe Ia), and
that of galaxy clustering data. We generalize the flux-averaging analysis
method of SNe Ia to allow correlated errors of SNe Ia, in order to reduce the
systematic bias due to weak lensing of SNe Ia. We find that flux-averaging
leads to larger errors on dark energy and cosmological parameters if only SN Ia
data are used. When SN Ia data (the latest compilation by the SNLS team) are
combined with WMAP 7 year results (in terms of our Gaussian fits to the
probability distributions of the CMB shift parameters), the latest Hubble
constant (H_0) measurement using the Hubble Space Telescope (HST), and gamma
ray burst (GRB) data, flux-averaging of SNe Ia increases the concordance with
other data, and leads to significantly tighter constraints on the dark energy
density at z=1, and the cosmic curvature \Omega_k. The galaxy clustering
measurements of H(z=0.35)r_s(z_d) and r_s(z_d)/D_A(z=0.35) (where H(z) is the
Hubble parameter, D_A(z) is the angular diameter distance, and r_s(z_d) is the
sound horizon at the drag epoch) by Chuang & Wang (2011) are consistent with SN
Ia data, given the same pirors (CMB+H_0+GRB), and lead to significantly
improved dark energy constraints when combined. Current data are fully
consistent with a cosmological constant and a flat universe.Comment: 11 pages, 9 figures. Slightly revised version, to appear in PRD.
Supernova flux-averaging code available at
http://www.nhn.ou.edu/~wang/SNcode
Life after extinction: palaeoenvironments of the earliest Triassic lower Katberg formation, including the origin of Lystrosaurus Bonebeds from the Karoo Basin, South Africa
Includes abstract.Includes bibliographical references.This study argues for the formation of bonebeds by aggregation of sub-adult Lystrosaurus during extended episodic periods of extreme climatic conditions, such as cold or drought, in the earliest Triassic Karoo Basin
- …