5,127 research outputs found

    A new development cycle of the Statistical Toolkit

    Full text link
    The Statistical Toolkit is an open source system specialized in the statistical comparison of distributions. It addresses requirements common to different experimental domains, such as simulation validation (e.g. comparison of experimental and simulated distributions), regression testing in the course of the software development process, and detector performance monitoring. Various sets of statistical tests have been added to the existing collection to deal with the one sample problem (i.e. the comparison of a data distribution to a function, including tests for normality, categorical analysis and the estimate of randomness). Improved algorithms and software design contribute to the robustness of the results. A simple user layer dealing with primitive data types facilitates the use of the toolkit both in standalone analyses and in large scale experiments.Comment: To be published in the Proc. of CHEP (Computing in High Energy Physics) 201

    Bayesian reweighting of nuclear PDFs and constraints from proton-lead collisions at the LHC

    Get PDF
    New hard-scattering measurements from the LHC proton-lead run have the potential to provide important constraints on the nuclear parton distributions and thus contributing to a better understanding of the initial state in heavy ion collisions. In order to quantify these constraints, as well as to assess the compatibility with available nuclear data from fixed target experiments and from RHIC, the traditional strategy is to perform a global fit of nuclear PDFs. This procedure is however time consuming and technically challenging, and moreover can only be performed by the PDF fitters themselves. In the case of proton PDFs, an alternative approach has been suggested that uses Bayesian inference to propagate the effects of new data into the PDFs without the need of refitting. In this work, we apply this reweighting procedure to study the impact on nuclear PDFs of low-mass Drell-Yan and single-inclusive hadroproduction pseudo-data from proton-lead collisions at the LHC as representative examples. In the hadroproduction case, in addition we assess the possibility of discriminating between the DGLAP and CGC production frameworks. We find that the LHC proton-lead data could lead to a substantial reduction of the uncertainties on nuclear PDFs, in particular for the small-x gluon PDF where uncertainties could decrease by up to a factor two. The Monte Carlo replicas of EPS09 used in the analysis are released as a public code for general use. It can be directly used, in particular, by the experimental collaborations to check, in a straightforward manner, the degree of compatibility of the new data with the global nPDF analyses.Comment: 21 pages, 10 figure

    Distributional Dominance with Dirty Data

    Get PDF
    Distributional dominance criteria are commonly applied to draw welfare inferences about comparisons, but conclusions drawn from empirical implementations of dominance criteria may be influenced by data contamination. We examine a non-parametric approach to refining Lorenz-type comparisons and apply the technique to two important examples from the LIS data-base.Distributional dominance, Lorenz curve, robustness.

    Robust Lorenz Curves: A Semiparametric Approach

    Get PDF
    Lorenz curves and second-order dominance criteria are known to be sensitive to data contamination in the right tail of the distribution. We propose two ways of dealing with the problem: (1) Estimate Lorenz curves using parametric models for income distributions, and (2) Combine empirical estimation with a parametric (robust) estimation of the upper tail of the distribution using the Pareto model. Approach (2) is preferred because of its flexibility. Using simulations we show the dramatic effect of a few contaminated data on the Lorenz ranking and the performance of the robust approach (2). Statistical inference tools are also provided.Welfare dominance, Lorenz curve, Pareto model, M-estimators.

    Statistical Inference for Lorenz Curves with Censored Data

    Get PDF
    Lorenz curves and associated tools for ranking income distributions are commonly estimated on the assumption that full, unbiased samples are available. However, it is common to find income and wealth distributions that are routinely censored or trimmed. We derive the sampling distribution for a key family of statistics in the case where data have been modified in this fashion.Lorenz curve, sampling errors

    Modelling Lorenz Curves:robust and semi-parametric issues

    Get PDF
    Modelling Lorenz curves (LC) for stochastic dominance comparisons is central to the analysis of income distribution. It is conventional to use non-parametric statistics based on empirical income cumulants which are in the construction of LC and other related second-order dominance criteria. However, although attractive because of its simplicity and its apparent flexibility, this approach suffers from important drawbacks. While no assumptions need to be made regarding the data-generating process (income distribution model), the empirical LC can be very sensitive to data particularities, especially in the upper tail of the distribution. This robustness problem can lead in practice to 'wrong' interpretation of dominance orders. A possible remedy for this problem is the use of parametric or semi-parametric models for the datagenerating process and robust estimators to obtain parameter estimates. In this paper, we focus on the robust estimation of semi parametric LC and investigate issues such as sensitivity of LC estimators to data contamination (Cowell and Victoria-Feser 2002), trimmed LC (Cowell and Victoria-Feser 2006) and inference for trimmed LC (Cowell and Victoria-Feser 2003), robust semi-parametric estimation for LC (Cowell and Victoria-Feser 2007) selection of optimal thresholds for (robust) semi parametric modelling (Dupuis and Victoria-Feser 2006) and use both simulations and real data to illustrate these points.

    Interpreting Spectral Energy Distributions from Young Stellar Objects. I. A grid of 200,000 YSO model SEDs

    Get PDF
    We present a grid of radiation transfer models of axisymmetric young stellar objects (YSOs), covering a wide range of stellar masses (from 0.1Msun to 50Msun) and evolutionary stages (from the early envelope infall stage to the late disk-only stage). The grid consists of 20,000 YSO models, with spectral energy distributions (SEDs) and polarization spectra computed at ten viewing angles for each model, resulting in a total of 200,000 SEDs. [...]. These models are publicly available on a dedicated WWW server: http://www.astro.wisc.edu/protostars/ . In this paper we summarize the main features of our models, as well as the range of parameters explored. [...]. We examine the dependence of the spectral indices of the model SEDs on envelope accretion rate and disk mass. In addition, we show variations of spectral indices with stellar temperature, disk inner radius, and disk flaring power for a subset of disk-only models. We also examine how changing the wavelength range of data used to calculate spectral indices affects their values. We show sample color-color plots of the entire grid as well as simulated clusters at various distances with typical {\it Spitzer Space Telescope} sensitivities. We find that young embedded sources generally occupy a large region of color-color space due to inclination and stellar temperature effects. Disk sources occupy a smaller region of color-color space, but overlap substantially with the region occupied by embedded sources, especially in the near- and mid-IR. We identify regions in color-color space where our models indicate that only sources at a given evolutionary stage should lie. [...].Comment: 69 pages, 28 figures, Accepted for publication in ApJS. Preprint with full resolution figures available at http://www.astro.wisc.edu/protostars

    Validating delta-filters for resonant bar detectors of improved bandwidth foreseeing the future coincidence with interferometers

    Full text link
    The classical delta filters used in the current resonant bar experiments for detecting GW bursts are viable when the bandwidth of resonant bars is few Hz. In that case, the incoming GW burst is likely to be viewed as an impulsive signal in a very narrow frequency window. After making improvements in the read-out with new transducers and high sensitivity dc-SQUID, the Explorer-Nautilus have improved the bandwidth (20\sim 20 Hz) at the sensitivity level of 1020/Hz10^{-20}/\sqrt{Hz}. Thus, it is necessary to reassess this assumption of delta-like signals while building filters in the resonant bars as the filtered output crucially depends on the shape of the waveform. This is presented with an example of GW signals -- stellar quasi-normal modes, by estimating the loss in SNR and the error in the timing, when the GW signal is filtered with the delta filter as compared to the optimal filter.Comment: 7 pages, presented in Amaldi6, accepted for publication in Journal of Physics: Conference Serie

    A Comparative Study of Dark Energy Constraints from Current Observational Data

    Full text link
    We examine how dark energy constraints from current observational data depend on the analysis methods used: the analysis of Type Ia supernovae (SNe Ia), and that of galaxy clustering data. We generalize the flux-averaging analysis method of SNe Ia to allow correlated errors of SNe Ia, in order to reduce the systematic bias due to weak lensing of SNe Ia. We find that flux-averaging leads to larger errors on dark energy and cosmological parameters if only SN Ia data are used. When SN Ia data (the latest compilation by the SNLS team) are combined with WMAP 7 year results (in terms of our Gaussian fits to the probability distributions of the CMB shift parameters), the latest Hubble constant (H_0) measurement using the Hubble Space Telescope (HST), and gamma ray burst (GRB) data, flux-averaging of SNe Ia increases the concordance with other data, and leads to significantly tighter constraints on the dark energy density at z=1, and the cosmic curvature \Omega_k. The galaxy clustering measurements of H(z=0.35)r_s(z_d) and r_s(z_d)/D_A(z=0.35) (where H(z) is the Hubble parameter, D_A(z) is the angular diameter distance, and r_s(z_d) is the sound horizon at the drag epoch) by Chuang & Wang (2011) are consistent with SN Ia data, given the same pirors (CMB+H_0+GRB), and lead to significantly improved dark energy constraints when combined. Current data are fully consistent with a cosmological constant and a flat universe.Comment: 11 pages, 9 figures. Slightly revised version, to appear in PRD. Supernova flux-averaging code available at http://www.nhn.ou.edu/~wang/SNcode

    Model-Independent Constraints on Dark Energy Density from Flux-averaging Analysis of Type Ia Supernova Data

    Get PDF
    We reconstruct the dark energy density ρX(z)\rho_X(z) as a free function from current type Ia supernova (SN Ia) data (Tonry et al. 2003; Barris et al. 2003; Knop et al. 2003), together with the Cosmic Microwave Background (CMB) shift parameter from CMB data (WMAP, CBI, and ACBAR), and the large scale structure (LSS) growth factor from 2dF galaxy survey data. We parametrize ρX(z)\rho_X(z) as a continuous function, given by interpolating its amplitudes at equally spaced zz values in the redshift range covered by SN Ia data, and a constant at larger zz (where ρX(z)\rho_X(z) is only weakly constrained by CMB data). We assume a flat universe, and use the Markov Chain Monte Carlo (MCMC) technique in our analysis. We find that the dark energy density ρX(z)\rho_X(z) is constant for 0 \la z \la 0.5 and increases with redshift zz for 0.5 \la z \la 1 at 68.3% confidence level, but is consistent with a constant at 95% confidence level. For comparison, we also give constraints on a constant equation of state for the dark energy. Flux-averaging of SN Ia data is required to yield cosmological parameter constraints that are free of the bias induced by weak gravitational lensing \citep{Wang00b}. We set up a consistent framework for flux-averaging analysis of SN Ia data, based on \cite{Wang00b}. We find that flux-averaging of SN Ia data leads to slightly lower Ωm\Omega_m and smaller time-variation in ρX(z)\rho_X(z). This suggests that a significant increase in the number of SNe Ia from deep SN surveys on a dedicated telescope \citep{Wang00a} is needed to place a robust constraint on the time-dependence of the dark energy density.Comment: Slightly revised in presentation, ApJ accepted. One color figure shows rho_X(z) reconstructed from dat
    corecore