2,454 research outputs found
An Application Specific Processor for Montecarlo Simulations
none6Proceedings editati da IEEEopenG. DANESE; M. BERA; F. LEPORATI; M. GIACHERO; N. NAZZICARI; A. SPELGATTIDanese, Giovanni; Bera, Marco; Leporati, Francesco; Giachero, Mauro; Nazzicari, NELSON DAVIDE; Spelgatti, Alvar
SIRENA: A CAD environment for behavioural modelling and simulation of VLSI cellular neural network chips
This paper presents SIRENA, a CAD environment for the simulation and modelling of mixed-signal VLSI parallel processing chips based on cellular neural networks. SIRENA includes capabilities for: (a) the description of nominal and non-ideal operation of CNN analogue circuitry at the behavioural level; (b) performing realistic simulations of the transient evolution of physical CNNs including deviations due to second-order effects of the hardware; and, (c) evaluating sensitivity figures, and realize noise and Monte Carlo simulations in the time domain. These capabilities portray SIRENA as better suited for CNN chip development than algorithmic simulation packages (such as OpenSimulator, Sesame) or conventional neural networks simulators (RCS, GENESIS, SFINX), which are not oriented to the evaluation of hardware non-idealities. As compared to conventional electrical simulators (such as HSPICE or ELDO-FAS), SIRENA provides easier modelling of the hardware parasitics, a significant reduction in computation time, and similar accuracy levels. Consequently, iteration during the design procedure becomes possible, supporting decision making regarding design strategies and dimensioning. SIRENA has been developed using object-oriented programming techniques in C, and currently runs under the UNIX operating system and X-Windows framework. It employs a dedicated high-level hardware description language: DECEL, fitted to the description of non-idealities arising in CNN hardware. This language has been developed aiming generality, in the sense of making no restrictions on the network models that can be implemented. SIRENA is highly modular and composed of independent tools. This simplifies future expansions and improvements.Comisión Interministerial de Ciencia y Tecnología TIC96-1392-C02-0
Unquenched QCD with Light Quarks
We present recent results in unquenched lattice QCD with two degenerate light
sea quarks using the truncated determinant approximation (TDA). In the TDA the
infrared modes contributing to the quark determinant are computed exactly up to
some cutoff in quark off-shellness (typically 2). This approach
allows simulations to be performed at much lighter quark masses than possible
with conventional hybrid MonteCarlo techniques. Results for the static energy
and topological charge distributions are presented using a large ensemble
generated on very coarse (6) but physically large lattices. Preliminary
results are also reported for the static energy and meson spectrum on 10x20
lattices (lattice scale =1.15 GeV) at quark masses corresponding to
pions of mass 200 MeV. Using multiboson simulation to compute the
ultraviolet part of the quark determinant the TDA approach becomes an exact
with essentially no increase in computational effort. Some preliminary results
using this fully unquenched algorithm are presented.Comment: LateX, 39 pages, 16 eps figures, 1 ps figur
Recommended from our members
Modelling the Spread of Botnet Malware in IoT-Based Wireless Sensor Networks
The propagation approach of a botnet largely dictates its formation, establishing a foundation of bots for future exploitation. The chosen propagation method determines the attack surface, and consequently, the degree of network penetration, as well as the overall size and the eventual attack potency. It is therefore essential to understand propagation behaviours and influential factors in order to better secure vulnerable systems. Whilst botnet propagation is generally well-studied, newer technologies like IoT have unique characteristics which are yet to be thoroughly explored. In this paper, we apply the principles of epidemic modelling to IoT networks consisting of wireless sensor nodes. We build IoT-SIS, a novel propagation model which considers the impact of IoT-specific characteristics like limited processing power, energy restrictions, and node density on the formation of a botnet. Focusing on worm-based propagation, this model is used to explore the dynamics of spread using numerical simulations and the Monte Carlo method, and to discuss the real-life implications of our findings
Improved Pseudofermion Approach for All-Point Propagators
Quark propagators with arbitrary sources and sinks can be obtained more
efficiently using a pseudofermion method with a mode-shifted action.
Mode-shifting solves the problem of critical slowing down (for light quarks)
induced by low eigenmodes of the Dirac operator. The method allows the full
physical content of every gauge configuration to be extracted, and should be
especially helpful for unquenched QCD calculations. The method can be applied
for all the conventional quark actions: Wilson, Sheikoleslami-Wohlert,
Kogut-Susskind, as well as Ginsparg-Wilson compliant overlap actions. The
statistical properties of the method are examined and examples of physical
processes under study are presented.Comment: LateX, 26 pages, 10 eps figure
Benchmarking Quantum Processor Performance at Scale
As quantum processors grow, new performance benchmarks are required to
capture the full quality of the devices at scale. While quantum volume is an
excellent benchmark, it focuses on the highest quality subset of the device and
so is unable to indicate the average performance over a large number of
connected qubits. Furthermore, it is a discrete pass/fail and so is not
reflective of continuous improvements in hardware nor does it provide
quantitative direction to large-scale algorithms. For example, there may be
value in error mitigated Hamiltonian simulation at scale with devices unable to
pass strict quantum volume tests. Here we discuss a scalable benchmark which
measures the fidelity of a connecting set of two-qubit gates over qubits by
measuring gate errors using simultaneous direct randomized benchmarking in
disjoint layers. Our layer fidelity can be easily related to algorithmic run
time, via defined in Ref.\cite{berg2022probabilistic} that can be used
to estimate the number of circuits required for error mitigation. The protocol
is efficient and obtains all the pair rates in the layered structure. Compared
to regular (isolated) RB this approach is sensitive to crosstalk. As an example
we measure a qubit layer fidelity on a 127 qubit fixed-coupling
"Eagle" processor (ibm\_sherbrooke) of 0.26(0.19) and on the 133 qubit
tunable-coupling "Heron" processor (ibm\_montecarlo) of 0.61(0.26). This can
easily be expressed as a layer size independent quantity, error per layered
gate (EPLG), which is here for
ibm\_sherbrooke and for ibm\_montecarlo.Comment: 15 pages, 8 figures (including appendices
Technique of uncertainty and sensitivity analysis for sustainable building energy systems performance calculations
Sustainable buildings design process is typical for modeling and simulation usage. The main reason is because there is generally no experience with such buildings and there is lot of new approaches and technical solutions to be used. Computer simulation could be supporting tool in engineering design process and can bring the good way for reducing energy consumption together with optimalization algorithm. For the optimization process we have to know which most sensitive input parametr from many of them has to be investigate. Therefore at first is necessary to perform the sensitivity analysis and find out the "strongest" input parametrs which most affecting the results under observation. Also still the simulation tools are mainly using to predict energy consumption, boiler and chiller loads, indoor air quality, etc. before the building is build. The information about the building envelope, schedule and HVAC components are unclear and can bring large uncertainty in results by setting this inputs to the simulation tools. Paper presents preview of uncertainty and sensitivity analysis. This techniques are shown on case study concretely BESTEST case600 with DRYCOLD climate conditions. Also systems VAV (variable volume of air) and water fancoil system are compared. For this prototype the simulation tool IES was chosen
Monte Carlo and Depletion Reactor Analysis for High-Performance Computing Applications
This dissertation discusses the research and development for a coupled neutron trans- port/isotopic depletion capability for use in high-preformance computing applications. Accurate neutronics modeling and simulation for \real reactor problems has been a long sought after goal in the computational community. A complementary \stretch goal to this is the ability to perform full-core depletion analysis and spent fuel isotopic characterization. This dissertation thus presents the research and development of a coupled Monte Carlo transport/isotopic depletion implementation with the Exnihilo framework geared for high-performance computing architectures to enable neutronics analysis for full-core reactor problems.
An in-depth case study of the current state of Monte Carlo neutron transport with respect to source sampling, source convergence, uncertainty underprediction and biases associated with localized tallies in Monte Carlo eigenvalue calculations was performed using MCNPand KENO. This analysis is utilized in the design and development of the statistical algorithms for Exnihilo\u27s Monte Carlo framework, Shift. To this end, a methodology has been developed in order to perform tally statistics in domain decomposed environments. This methodology has been shown to produce accurate tally uncertainty estimates in domain-decomposed environments without a significant increase in the memory requirements, processor-to-processor communications, or computational biases.
With the addition of parallel, domain-decomposed tally uncertainty estimation processes, a depletion package was developed for the Exnihilo code suite to utilize the depletion capabilities of the Oak Ridge Isotope GENeration code. This interface was designed to be transport agnostic, meaning that it can be used by any of the reactor analysis packages within Exnihilo such as Denovo or Shift. Extensive validation and testing of the ORIGEN interface and coupling with the Shift Monte Carlo transport code is performed within this dissertation, and results are presented for the calculated eigenvalues, material powers, and nuclide concentrations for the depleted materials. These results are then compared to ORIGEN and TRITON depletion calculations, and analysis shows that the Exnihilo transport-depletion capability is in good agreement with these codes
Design and optimization of a portable LQCD Monte Carlo code using OpenACC
The present panorama of HPC architectures is extremely heterogeneous, ranging
from traditional multi-core CPU processors, supporting a wide class of
applications but delivering moderate computing performance, to many-core GPUs,
exploiting aggressive data-parallelism and delivering higher performances for
streaming computing applications. In this scenario, code portability (and
performance portability) become necessary for easy maintainability of
applications; this is very relevant in scientific computing where code changes
are very frequent, making it tedious and prone to error to keep different code
versions aligned. In this work we present the design and optimization of a
state-of-the-art production-level LQCD Monte Carlo application, using the
directive-based OpenACC programming model. OpenACC abstracts parallel
programming to a descriptive level, relieving programmers from specifying how
codes should be mapped onto the target architecture. We describe the
implementation of a code fully written in OpenACC, and show that we are able to
target several different architectures, including state-of-the-art traditional
CPUs and GPUs, with the same code. We also measure performance, evaluating the
computing efficiency of our OpenACC code on several architectures, comparing
with GPU-specific implementations and showing that a good level of
performance-portability can be reached.Comment: 26 pages, 2 png figures, preprint of an article submitted for
consideration in International Journal of Modern Physics
- …