247 research outputs found

    HP-CERTI: Towards a high performance, high availability open source RTI for composable simulations (04F-SIW-014)

    Get PDF
    Composing simulations of complex systems from already existing simulation components remains a challenging issue. Motivations for composable simulation include generation of a given federation driven by operational requirements provided "on the fly". The High Level Architecture, initially developed for designing fully distributed simulations, can be considered as an interoperability standard for composing simulations from existing components. Requirements for constructing such complex simulations are quite different from those discussed for distributed simulations. Although interoperability and reusability remain essential, both high performance and availability have also to be considered to fulfill the requirements of the end user. ONERA is currently designing a High Performance / High Availability HLA Run-time Infrastructure from its open source implementation of HLA 1.3 specifications. HP-CERTI is a software package including two main components: the first one, SHM-CERTI, provides an optimized version of CERTI based on a shared memory communication scheme; the second one, Kerrighed-CERTI, allows the deployment of CERTI through the control of the Kerrighed Single System Image operating system for clusters, currently designed by IRISA. This paper describes the design of both high performance and availability Runtime Infrastructures, focusing on the architecture of SHM-CERTI. This work is carried out in the context of the COCA (High Performance Distributed Simulation and Models Reuse) Project, sponsored by the DGA/STTC (Délégation Générale pour l'Armement/Service des Stratégies Techniques et des Technologies Communes) of the French Ministry of Defense

    Adaptive Probabilistic Flooding for Multipath Routing

    Full text link
    In this work, we develop a distributed source routing algorithm for topology discovery suitable for ISP transport networks, that is however inspired by opportunistic algorithms used in ad hoc wireless networks. We propose a plug-and-play control plane, able to find multiple paths toward the same destination, and introduce a novel algorithm, called adaptive probabilistic flooding, to achieve this goal. By keeping a small amount of state in routers taking part in the discovery process, our technique significantly limits the amount of control messages exchanged with flooding -- and, at the same time, it only minimally affects the quality of the discovered multiple path with respect to the optimal solution. Simple analytical bounds, confirmed by results gathered with extensive simulation on four realistic topologies, show our approach to be of high practical interest.Comment: 6 pages, 6 figure

    Improved cosmological constraints from a joint analysis of the SDSS-II and SNLS supernova samples

    Get PDF
    Aims. We present cosmological constraints from a joint analysis of type Ia supernova (SN Ia) observations obtained by the SDSS-II and SNLS collaborations. The dataset includes several low-redshift samples (z< 0.1), all three seasons from the SDSS-II (0.05 <z< 0.4), and three years from SNLS (0.2 <z< 1), and it totals 740 spectroscopically confirmed type Ia supernovae with high-quality light curves. Methods. We followed the methods and assumptions of the SNLS three-year data analysis except for the following important improvements: 1) the addition of the full SDSS-II spectroscopically-confirmed SN Ia sample in both the training of the SALT2 light-curve model and in the Hubble diagram analysis (374 SNe); 2) intercalibration of the SNLS and SDSS surveys and reduced systematic uncertainties in the photometric calibration, performed blindly with respect to the cosmology analysis; and 3) a thorough investigation of systematic errors associated with the SALT2 modeling of SN Ia light curves. Results. We produce recalibrated SN Ia light curves and associated distances for the SDSS-II and SNLS samples. The large SDSS-II sample provides an effective, independent, low-z anchor for the Hubble diagram and reduces the systematic error from calibration systematics in the low-z SN sample. For a flat ΛCDM cosmology, we find Ωm =0.295 ± 0.034 (stat+sys), a value consistent with the most recent cosmic microwave background (CMB) measurement from the Planck and WMAP experiments. Our result is 1.8σ (stat+sys) different than the previously published result of SNLS three-year data. The change is due primarily to improvements in the SNLS photometric calibration. When combined with CMB constraints, we measure a constant dark-energy equation of state parameter w =−1.018 ± 0.057 (stat+sys) for a flat universe. Adding baryon acoustic oscillation distance measurements gives similar constraints: w =−1.027 ± 0.055. Our supernova measurements provide the most stringent constraints to date on the nature of dark energy

    Photometry of supernovae in an image series : methods and application to the Supernova Legacy Survey (SNLS)

    Full text link
    We present a technique to measure lightcurves of time-variable point sources on a spatially structured background from imaging data. The technique was developed to measure light curves of SNLS supernovae in order to infer their distances. This photometry technique performs simultaneous PSF photometry at the same sky position on an image series. We describe two implementations of the method: one that resamples images before measuring fluxes, and one which does not. In both instances, we sketch the key algorithms involved and present the validation using semi-artificial sources introduced in real images in order to assess the accuracy of the supernova flux measurements relative to that of surrounding stars. We describe the methods required to anchor these PSF fluxes to calibrated aperture catalogs, in order to derive SN magnitudes. We find a marginally significant bias of 2 mmag of the after-resampling method, and no bias at the mmag accuracy for the non-resampling method. Given surrounding star magnitudes, we determine the systematic uncertainty of SN magnitudes to be less than 1.5 mmag, which represents about one third of the current photometric calibration uncertainty affecting SN measurements. The SN photometry delivers several by-products: bright star PSF flux mea- surements which have a repeatability of about 0.6%, as for aperture measurements; we measure relative astrometric positions with a noise floor of 2.4 mas for a single-image bright star measurement; we show that in all bands of the MegaCam instrument, stars exhibit a profile linearly broadening with flux by about 0.5% over the whole brightness range.Comment: Accepted for publication in A&A. 20 page

    Forecast B-modes detection at large scales in presence of noise and foregrounds

    Full text link
    We investigate the detectability of the primordial CMB polarization B-mode power spectrum on large scales in the presence of instrumental noise and realistic foreground contamination. We have worked out a method to estimate the errors on component separation and to propagate them up to the power spectrum estimation. The performances of our method are illustrated by applying it to the instrumental specifications of the Planck satellite and to the proposed configuration for the next generation CMB polarization experiment COrE. We demonstrate that a proper component separation step is required in order achieve the detection of B-modes on large scales and that the final sensitivity to B-modes of a given experiment is determined by a delicate balance between noise level and residual foregrounds, which depend on the set of frequencies exploited in the CMB reconstruction, on the signal-to-noise of each frequency map, and on our ability to correctly model the spectral behavior of the foreground components. We have produced a flexible software tool that allows the comparison of performances on B-mode detection of different instrumental specifications (choice of frequencies, noise level at each frequency, etc.) as well as of different proposed approaches to component separation.Comment: 7 pages, 2 tables, 1 figure, accepted by MNRA

    The SkyMapper Transient Survey

    Full text link
    The SkyMapper 1.3 m telescope at Siding Spring Observatory has now begun regular operations. Alongside the Southern Sky Survey, a comprehensive digital survey of the entire southern sky, SkyMapper will carry out a search for supernovae and other transients. The search strategy, covering a total footprint area of ~2000 deg2 with a cadence of ≀5\leq 5 days, is optimised for discovery and follow-up of low-redshift type Ia supernovae to constrain cosmic expansion and peculiar velocities. We describe the search operations and infrastructure, including a parallelised software pipeline to discover variable objects in difference imaging; simulations of the performance of the survey over its lifetime; public access to discovered transients; and some first results from the Science Verification data.Comment: 13 pages, 11 figures; submitted to PAS

    The DICE calibration project: design, characterization, and first results

    Full text link
    We describe the design, operation, and first results of a photometric calibration project, called DICE (Direct Illumination Calibration Experiment), aiming at achieving precise instrumental calibration of optical telescopes. The heart of DICE is an illumination device composed of 24 narrow-spectrum, high-intensity, light-emitting diodes (LED) chosen to cover the ultraviolet-to-near-infrared spectral range. It implements a point-like source placed at a finite distance from the telescope entrance pupil, yielding a flat field illumination that covers the entire field of view of the imager. The purpose of this system is to perform a lightweight routine monitoring of the imager passbands with a precision better than 5 per-mil on the relative passband normalisations and about 3{\AA} on the filter cutoff positions. The light source is calibrated on a spectrophotometric bench. As our fundamental metrology standard, we use a photodiode calibrated at NIST. The radiant intensity of each beam is mapped, and spectra are measured for each LED. All measurements are conducted at temperatures ranging from 0{\deg}C to 25{\deg}C in order to study the temperature dependence of the system. The photometric and spectroscopic measurements are combined into a model that predicts the spectral intensity of the source as a function of temperature. We find that the calibration beams are stable at the 10−410^{-4} level -- after taking the slight temperature dependence of the LED emission properties into account. We show that the spectral intensity of the source can be characterised with a precision of 3{\AA} in wavelength. In flux, we reach an accuracy of about 0.2-0.5% depending on how we understand the off-diagonal terms of the error budget affecting the calibration of the NIST photodiode. With a routine 60-mn calibration program, the apparatus is able to constrain the passbands at the targeted precision levels.Comment: 25 pages, 27 figures, accepted for publication in A&
    • 

    corecore