5,531 research outputs found
High performance computing of explicit schemes for electrofusion jointing process based on message-passing paradigm
The research focused on heterogeneous cluster workstations comprising of a number of CPUs in single and shared architecture platform. The problem statements under consideration involved one dimensional parabolic equations. The thermal process of electrofusion jointing was also discussed. Numerical schemes of explicit type such as AGE, Brian, and Charlies Methods were employed. The parallelization of these methods were based on the domain decomposition technique. Some parallel performance measurement for these methods were also addressed. Temperature profile of the one dimensional radial model of the electrofusion process were also given
QTM: computational package using MPI protocol for quantum trajectories method
The Quantum Trajectories Method (QTM) is one of {the} frequently used methods
for studying open quantum systems. { The main idea of this method is {the}
evolution of wave functions which {describe the system (as functions of time).
Then,} so-called quantum jumps are applied at {a} randomly selected point in
time. {The} obtained system state is called as a trajectory. After averaging
many single trajectories{,} we obtain the approximation of the behavior of {a}
quantum system.} {This fact also allows} us to use parallel computation
methods. In the article{,} we discuss the QTM package which is supported by the
MPI technology. Using MPI allowed {utilizing} the parallel computing for
calculating the trajectories and averaging them -- as the effect of these
actions{,} the time {taken by} calculations is shorter. In spite of using the
C++ programming language, the presented solution is easy to utilize and does
not need any advanced programming techniques. At the same time{,} it offers a
higher performance than other packages realizing the QTM. It is especially
important in the case of harder computational tasks{,} and the use of MPI
allows {improving the} performance of particular problems which can be solved
in the field of open quantum systems.Comment: 28 pages, 9 figure
Recommended from our members
Leveraging legacy codes to distributed problem solving environments: A web service approach
This paper describes techniques used to leverage high performance legacy codes as CORBA components to a distributed problem solving environment. It first briefly introduces the software architecture adopted by the environment. Then it presents a CORBA oriented wrapper generator (COWG) which can be used to automatically wrap high performance legacy codes as CORBA components. Two legacy codes have been wrapped with COWG. One is an MPI-based molecular dynamic simulation (MDS) code, the other is a finite element based computational fluid dynamics (CFD) code for simulating incompressible Navier-Stokes flows. Performance comparisons between runs of the MDS CORBA component and the original MDS legacy code on a cluster of workstations and on a parallel computer are also presented. Wrapped as CORBA components, these legacy codes can be reused in a distributed computing environment. The first case shows that high performance can be maintained with the wrapped MDS component. The second case shows that a Web user can submit a task to the wrapped CFD component through a Web page without knowing the exact implementation of the component. In this way, a userâs desktop computing environment can be extended to a high performance computing environment using a cluster of workstations or a parallel computer
On the parallelization of stellar evolution codes
Multidimensional nucleosynthesis studies with hundreds of nuclei linked through thousands of nuclear processes are still computationally prohibitive. To date, most nucleosynthesis studies rely either on hydrostatic/hydrodynamic simulations in spherical symmetry, or on post-processing simulations using temperature and density versus time profiles directly linked to huge nuclear reaction networks.
Parallel computing has been regarded as the main permitting factor of computationally intensive simulations. This paper explores the different pros and cons in the parallelization of stellar codes, providing recommendations on when and how parallelization may help in improving the performance of a code for
astrophysical applications.
We report on different parallelization strategies succesfully applied to the spherically symmetric, Lagrangian, implicit hydrodynamic code SHIVA, extensively used in the modeling of classical novae and type I X-ray bursts.
When only matrix build-up and inversion processes in the nucleosynthesis subroutines are parallelized (a suitable approach for post-processing calculations), the huge amount of time spent on communications between cores, together with the small problem size (limited by the number of isotopes of the nuclear
network), result in a much worse performance of the parallel application than the 1-core, sequential version of the code. Parallelization of the matrix build-up and inversion processes in the nucleosynthesis subroutines is not recommended unless the number of isotopes adopted largely exceeds 10,000.
In sharp contrast, speed-up factors of 26 and 35 have been obtained with a parallelized version of
SHIVA, in a 200-shell simulation of a type I X-ray burstcarried out with two nuclear reaction networks: a reduced one, consisting of 324 isotopes and 1392 reactions, and a more extended network with 60
6 nuclides and 3551 nuclear interactions. Maximum speed-ups of ~41 (324-isotope network) and ~85 (606-isotope network), are also predicted for 200 cores, stressing that the number of shells of the computational domain constitutes an effective upper limit for the maximum number of cores that could be used in a parallel
application.Peer ReviewedPostprint (published version
Turbomachinery CFD on parallel computers
The role of multistage turbomachinery simulation in the development of propulsion system models is discussed. Particularly, the need for simulations with higher fidelity and faster turnaround time is highlighted. It is shown how such fast simulations can be used in engineering-oriented environments. The use of parallel processing to achieve the required turnaround times is discussed. Current work by several researchers in this area is summarized. Parallel turbomachinery CFD research at the NASA Lewis Research Center is then highlighted. These efforts are focused on implementing the average-passage turbomachinery model on MIMD, distributed memory parallel computers. Performance results are given for inviscid, single blade row and viscous, multistage applications on several parallel computers, including networked workstations
Recommended from our members
Domain decomposition based algorithms for some inverse problems
The work presented in this thesis develop algorithms to solve inverse problems where source terms are unknown. The algorithms are developed 011frameworks provided by domain decomposition methods and the numerical schemes use finite volume and finite difference discretisations.
Three algorithms are developed in the context of a metal cutting problem. The algorithms require measurement data within the physical body in order to retrieve the temperature field and the unknown source terms. It is shown that the algorithms can retrieve both the temperature field and the unknown source accurately. Applicability of the algorithms to other problems is shown by using one of the algorithms to solve a welding problem.
Presence of untreated noisy measurement data can severely affect the accuracy of the retrieved source. It is illustrated that a simple noise treatment procedure such as a least squares method can remedy this situation. The algorithms are implemented 011parallel computing platforms to reduce the execution time. By exploiting domain and data parallelism within the algorithms significant performance improvements are achieved. It is also shown that by exploiting mathematical properties such as change of nonlinearity further performance improvements can be made
- âŠ