2,454 research outputs found

    An Application Specific Processor for Montecarlo Simulations

    Get PDF
    none6Proceedings editati da IEEEopenG. DANESE; M. BERA; F. LEPORATI; M. GIACHERO; N. NAZZICARI; A. SPELGATTIDanese, Giovanni; Bera, Marco; Leporati, Francesco; Giachero, Mauro; Nazzicari, NELSON DAVIDE; Spelgatti, Alvar

    SIRENA: A CAD environment for behavioural modelling and simulation of VLSI cellular neural network chips

    Get PDF
    This paper presents SIRENA, a CAD environment for the simulation and modelling of mixed-signal VLSI parallel processing chips based on cellular neural networks. SIRENA includes capabilities for: (a) the description of nominal and non-ideal operation of CNN analogue circuitry at the behavioural level; (b) performing realistic simulations of the transient evolution of physical CNNs including deviations due to second-order effects of the hardware; and, (c) evaluating sensitivity figures, and realize noise and Monte Carlo simulations in the time domain. These capabilities portray SIRENA as better suited for CNN chip development than algorithmic simulation packages (such as OpenSimulator, Sesame) or conventional neural networks simulators (RCS, GENESIS, SFINX), which are not oriented to the evaluation of hardware non-idealities. As compared to conventional electrical simulators (such as HSPICE or ELDO-FAS), SIRENA provides easier modelling of the hardware parasitics, a significant reduction in computation time, and similar accuracy levels. Consequently, iteration during the design procedure becomes possible, supporting decision making regarding design strategies and dimensioning. SIRENA has been developed using object-oriented programming techniques in C, and currently runs under the UNIX operating system and X-Windows framework. It employs a dedicated high-level hardware description language: DECEL, fitted to the description of non-idealities arising in CNN hardware. This language has been developed aiming generality, in the sense of making no restrictions on the network models that can be implemented. SIRENA is highly modular and composed of independent tools. This simplifies future expansions and improvements.Comisión Interministerial de Ciencia y Tecnología TIC96-1392-C02-0

    Unquenched QCD with Light Quarks

    Get PDF
    We present recent results in unquenched lattice QCD with two degenerate light sea quarks using the truncated determinant approximation (TDA). In the TDA the infrared modes contributing to the quark determinant are computed exactly up to some cutoff in quark off-shellness (typically 2ΛQCD\Lambda_{QCD}). This approach allows simulations to be performed at much lighter quark masses than possible with conventional hybrid MonteCarlo techniques. Results for the static energy and topological charge distributions are presented using a large ensemble generated on very coarse (64^4) but physically large lattices. Preliminary results are also reported for the static energy and meson spectrum on 103^3x20 lattices (lattice scale a1a^{-1}=1.15 GeV) at quark masses corresponding to pions of mass \leq 200 MeV. Using multiboson simulation to compute the ultraviolet part of the quark determinant the TDA approach becomes an exact with essentially no increase in computational effort. Some preliminary results using this fully unquenched algorithm are presented.Comment: LateX, 39 pages, 16 eps figures, 1 ps figur

    Improved Pseudofermion Approach for All-Point Propagators

    Get PDF
    Quark propagators with arbitrary sources and sinks can be obtained more efficiently using a pseudofermion method with a mode-shifted action. Mode-shifting solves the problem of critical slowing down (for light quarks) induced by low eigenmodes of the Dirac operator. The method allows the full physical content of every gauge configuration to be extracted, and should be especially helpful for unquenched QCD calculations. The method can be applied for all the conventional quark actions: Wilson, Sheikoleslami-Wohlert, Kogut-Susskind, as well as Ginsparg-Wilson compliant overlap actions. The statistical properties of the method are examined and examples of physical processes under study are presented.Comment: LateX, 26 pages, 10 eps figure

    Benchmarking Quantum Processor Performance at Scale

    Full text link
    As quantum processors grow, new performance benchmarks are required to capture the full quality of the devices at scale. While quantum volume is an excellent benchmark, it focuses on the highest quality subset of the device and so is unable to indicate the average performance over a large number of connected qubits. Furthermore, it is a discrete pass/fail and so is not reflective of continuous improvements in hardware nor does it provide quantitative direction to large-scale algorithms. For example, there may be value in error mitigated Hamiltonian simulation at scale with devices unable to pass strict quantum volume tests. Here we discuss a scalable benchmark which measures the fidelity of a connecting set of two-qubit gates over NN qubits by measuring gate errors using simultaneous direct randomized benchmarking in disjoint layers. Our layer fidelity can be easily related to algorithmic run time, via γ\gamma defined in Ref.\cite{berg2022probabilistic} that can be used to estimate the number of circuits required for error mitigation. The protocol is efficient and obtains all the pair rates in the layered structure. Compared to regular (isolated) RB this approach is sensitive to crosstalk. As an example we measure a N=80 (100)N=80~(100) qubit layer fidelity on a 127 qubit fixed-coupling "Eagle" processor (ibm\_sherbrooke) of 0.26(0.19) and on the 133 qubit tunable-coupling "Heron" processor (ibm\_montecarlo) of 0.61(0.26). This can easily be expressed as a layer size independent quantity, error per layered gate (EPLG), which is here 1.7×102(1.7×102)1.7\times10^{-2}(1.7\times10^{-2}) for ibm\_sherbrooke and 6.2×103(1.2×102)6.2\times10^{-3}(1.2\times10^{-2}) for ibm\_montecarlo.Comment: 15 pages, 8 figures (including appendices

    Implementation of Hardware Accelerators on Zynq

    Get PDF

    Technique of uncertainty and sensitivity analysis for sustainable building energy systems performance calculations

    Get PDF
    Sustainable buildings design process is typical for modeling and simulation usage. The main reason is because there is generally no experience with such buildings and there is lot of new approaches and technical solutions to be used. Computer simulation could be supporting tool in engineering design process and can bring the good way for reducing energy consumption together with optimalization algorithm. For the optimization process we have to know which most sensitive input parametr from many of them has to be investigate. Therefore at first is necessary to perform the sensitivity analysis and find out the "strongest" input parametrs which most affecting the results under observation. Also still the simulation tools are mainly using to predict energy consumption, boiler and chiller loads, indoor air quality, etc. before the building is build. The information about the building envelope, schedule and HVAC components are unclear and can bring large uncertainty in results by setting this inputs to the simulation tools. Paper presents preview of uncertainty and sensitivity analysis. This techniques are shown on case study concretely BESTEST case600 with DRYCOLD climate conditions. Also systems VAV (variable volume of air) and water fancoil system are compared. For this prototype the simulation tool IES was chosen

    Monte Carlo and Depletion Reactor Analysis for High-Performance Computing Applications

    Get PDF
    This dissertation discusses the research and development for a coupled neutron trans- port/isotopic depletion capability for use in high-preformance computing applications. Accurate neutronics modeling and simulation for \real reactor problems has been a long sought after goal in the computational community. A complementary \stretch goal to this is the ability to perform full-core depletion analysis and spent fuel isotopic characterization. This dissertation thus presents the research and development of a coupled Monte Carlo transport/isotopic depletion implementation with the Exnihilo framework geared for high-performance computing architectures to enable neutronics analysis for full-core reactor problems. An in-depth case study of the current state of Monte Carlo neutron transport with respect to source sampling, source convergence, uncertainty underprediction and biases associated with localized tallies in Monte Carlo eigenvalue calculations was performed using MCNPand KENO. This analysis is utilized in the design and development of the statistical algorithms for Exnihilo\u27s Monte Carlo framework, Shift. To this end, a methodology has been developed in order to perform tally statistics in domain decomposed environments. This methodology has been shown to produce accurate tally uncertainty estimates in domain-decomposed environments without a significant increase in the memory requirements, processor-to-processor communications, or computational biases. With the addition of parallel, domain-decomposed tally uncertainty estimation processes, a depletion package was developed for the Exnihilo code suite to utilize the depletion capabilities of the Oak Ridge Isotope GENeration code. This interface was designed to be transport agnostic, meaning that it can be used by any of the reactor analysis packages within Exnihilo such as Denovo or Shift. Extensive validation and testing of the ORIGEN interface and coupling with the Shift Monte Carlo transport code is performed within this dissertation, and results are presented for the calculated eigenvalues, material powers, and nuclide concentrations for the depleted materials. These results are then compared to ORIGEN and TRITON depletion calculations, and analysis shows that the Exnihilo transport-depletion capability is in good agreement with these codes

    Design and optimization of a portable LQCD Monte Carlo code using OpenACC

    Full text link
    The present panorama of HPC architectures is extremely heterogeneous, ranging from traditional multi-core CPU processors, supporting a wide class of applications but delivering moderate computing performance, to many-core GPUs, exploiting aggressive data-parallelism and delivering higher performances for streaming computing applications. In this scenario, code portability (and performance portability) become necessary for easy maintainability of applications; this is very relevant in scientific computing where code changes are very frequent, making it tedious and prone to error to keep different code versions aligned. In this work we present the design and optimization of a state-of-the-art production-level LQCD Monte Carlo application, using the directive-based OpenACC programming model. OpenACC abstracts parallel programming to a descriptive level, relieving programmers from specifying how codes should be mapped onto the target architecture. We describe the implementation of a code fully written in OpenACC, and show that we are able to target several different architectures, including state-of-the-art traditional CPUs and GPUs, with the same code. We also measure performance, evaluating the computing efficiency of our OpenACC code on several architectures, comparing with GPU-specific implementations and showing that a good level of performance-portability can be reached.Comment: 26 pages, 2 png figures, preprint of an article submitted for consideration in International Journal of Modern Physics
    corecore