347 research outputs found

    A Novel Generic Framework for Track Fitting in Complex Detector Systems

    Full text link
    This paper presents a novel framework for track fitting which is usable in a wide range of experiments, independent of the specific event topology, detector setup, or magnetic field arrangement. This goal is achieved through a completely modular design. Fitting algorithms are implemented as interchangeable modules. At present, the framework contains a validated Kalman filter. Track parameterizations and the routines required to extrapolate the track parameters and their covariance matrices through the experiment are also implemented as interchangeable modules. Different track parameterizations and extrapolation routines can be used simultaneously for fitting of the same physical track. Representations of detector hits are the third modular ingredient to the framework. The hit dimensionality and orientation of planar tracking detectors are not restricted. Tracking information from detectors which do not measure the passage of particles in a fixed physical detector plane, e.g. drift chambers or TPCs, is used without any simplifications. The concept is implemented in a light-weight C++ library called GENFIT, which is available as free software

    Synthetic LISA: Simulating Time Delay Interferometry in a Model LISA

    Full text link
    We report on three numerical experiments on the implementation of Time-Delay Interferometry (TDI) for LISA, performed with Synthetic LISA, a C++/Python package that we developed to simulate the LISA science process at the level of scientific and technical requirements. Specifically, we study the laser-noise residuals left by first-generation TDI when the LISA armlengths have a realistic time dependence; we characterize the armlength-measurements accuracies that are needed to have effective laser-noise cancellation in both first- and second-generation TDI; and we estimate the quantization and telemetry bitdepth needed for the phase measurements. Synthetic LISA generates synthetic time series of the LISA fundamental noises, as filtered through all the TDI observables; it also provides a streamlined module to compute the TDI responses to gravitational waves according to a full model of TDI, including the motion of the LISA array and the temporal and directional dependence of the armlengths. We discuss the theoretical model that underlies the simulation, its implementation, and its use in future investigations on system characterization and data-analysis prototyping for LISA.Comment: 18 pages, 14 EPS figures, REVTeX 4. Accepted PRD version. See http://www.vallis.org/syntheticlisa for information on the Synthetic LISA software packag

    Spectral Line Removal in the LIGO Data Analysis System (LDAS)

    Full text link
    High power in narrow frequency bands, spectral lines, are a feature of an interferometric gravitational wave detector's output. Some lines are coherent between interferometers, in particular, the 2 km and 4 km LIGO Hanford instruments. This is of concern to data analysis techniques, such as the stochastic background search, that use correlations between instruments to detect gravitational radiation. Several techniques of `line removal' have been proposed. Where a line is attributable to a measurable environmental disturbance, a simple linear model may be fitted to predict, and subsequently subtract away, that line. This technique has been implemented (as the command oelslr) in the LIGO Data Analysis System (LDAS). We demonstrate its application to LIGO S1 data.Comment: 11 pages, 5 figures, to be published in CQG GWDAW02 proceeding

    Logarithmic growth dynamics in software networks

    Full text link
    In a recent paper, Krapivsky and Redner (Phys. Rev. E, 71 (2005) 036118) proposed a new growing network model with new nodes being attached to a randomly selected node, as well to all ancestors of the target node. The model leads to a sparse graph with an average degree growing logarithmically with the system size. Here we present compeling evidence for software networks being the result of a similar class of growing dynamics. The predicted pattern of network growth, as well as the stationary in- and out-degree distributions are consistent with the model. Our results confirm the view of large-scale software topology being generated through duplication-rewiring mechanisms. Implications of these findings are outlined.Comment: 7 pages, 3 figures, published in Europhysics Letters (2005

    Validation of Kalman Filter alignment algorithm with cosmic-ray data using a CMS silicon strip tracker endcap

    Full text link
    A Kalman Filter alignment algorithm has been applied to cosmic-ray data. We discuss the alignment algorithm and an experiment-independent implementation including outlier rejection and treatment of weakly determined parameters. Using this implementation, the algorithm has been applied to data recorded with one CMS silicon tracker endcap. Results are compared to both photogrammetry measurements and data obtained from a dedicated hardware alignment system, and good agreement is observed.Comment: 11 pages, 8 figures. CMS NOTE-2010/00

    CMBEASY:: an Object Oriented Code for the Cosmic Microwave Background

    Full text link
    We have ported the cmbfast package to the C++ programming language to produce cmbeasy, an object oriented code for the cosmic microwave background. The code is available at www.cmbeasy.org. We sketch the design of the new code, emphasizing the benefits of object orientation in cosmology, which allow for simple substitution of different cosmological models and gauges. Both gauge invariant perturbations and quintessence support has been added to the code. For ease of use, as well as for instruction, a graphical user interface is available.Comment: 7 pages, 5 figures, matches published version, code at http://www.cmbeasy.or

    Towards a fully automated computation of RG-functions for the 3-dd O(N) vector model: Parametrizing amplitudes

    Full text link
    Within the framework of field-theoretical description of second-order phase transitions via the 3-dimensional O(N) vector model, accurate predictions for critical exponents can be obtained from (resummation of) the perturbative series of Renormalization-Group functions, which are in turn derived --following Parisi's approach-- from the expansions of appropriate field correlators evaluated at zero external momenta. Such a technique was fully exploited 30 years ago in two seminal works of Baker, Nickel, Green and Meiron, which lead to the knowledge of the β\beta-function up to the 6-loop level; they succeeded in obtaining a precise numerical evaluation of all needed Feynman amplitudes in momentum space by lowering the dimensionalities of each integration with a cleverly arranged set of computational simplifications. In fact, extending this computation is not straightforward, due both to the factorial proliferation of relevant diagrams and the increasing dimensionality of their associated integrals; in any case, this task can be reasonably carried on only in the framework of an automated environment. On the road towards the creation of such an environment, we here show how a strategy closely inspired by that of Nickel and coworkers can be stated in algorithmic form, and successfully implemented on the computer. As an application, we plot the minimized distributions of residual integrations for the sets of diagrams needed to obtain RG-functions to the full 7-loop level; they represent a good evaluation of the computational effort which will be required to improve the currently available estimates of critical exponents.Comment: 54 pages, 17 figures and 4 table

    The Verifying Compiler: A Grand Challenge for Computing Research

    Get PDF
    Abstract. This contribution proposes a set of criteria that distinguish a grand challenge in science or engineering from the many other kinds of short-term or long-term research problems that engage the interest of scientists and engineers. As an example drawn from Computer Science, it revives an old challenge: the construction and application of a verifying compiler that guarantees correctness of a program before running it. Introduction. The primary purpose of the formulation and promulgation of a grand challenge is the advancement of science or engineering. A grand challenge represents a commitment by a significant section of the research community to work together towards a common goal, agreed to be valuable and achievable by a team effort within a predicted timescale. The challenge is formulated by th
    corecore