10,914 research outputs found
Open-architecture Implementation of Fragment Molecular Orbital Method for Peta-scale Computing
We present our perspective and goals on highperformance computing for
nanoscience in accordance with the global trend toward "peta-scale computing."
After reviewing our results obtained through the grid-enabled version of the
fragment molecular orbital method (FMO) on the grid testbed by the Japanese
Grid Project, National Research Grid Initiative (NAREGI), we show that FMO is
one of the best candidates for peta-scale applications by predicting its
effective performance in peta-scale computers. Finally, we introduce our new
project constructing a peta-scale application in an open-architecture
implementation of FMO in order to realize both goals of highperformance in
peta-scale computers and extendibility to multiphysics simulations.Comment: 6 pages, 9 figures, proceedings of the 2nd IEEE/ACM international
workshop on high performance computing for nano-science and technology
(HPCNano06
Steering in computational science: mesoscale modelling and simulation
This paper outlines the benefits of computational steering for high
performance computing applications. Lattice-Boltzmann mesoscale fluid
simulations of binary and ternary amphiphilic fluids in two and three
dimensions are used to illustrate the substantial improvements which
computational steering offers in terms of resource efficiency and time to
discover new physics. We discuss details of our current steering
implementations and describe their future outlook with the advent of
computational grids.Comment: 40 pages, 11 figures. Accepted for publication in Contemporary
Physic
SchNet - a deep learning architecture for molecules and materials
Deep learning has led to a paradigm shift in artificial intelligence,
including web, text and image search, speech recognition, as well as
bioinformatics, with growing impact in chemical physics. Machine learning in
general and deep learning in particular is ideally suited for representing
quantum-mechanical interactions, enabling to model nonlinear potential-energy
surfaces or enhancing the exploration of chemical compound space. Here we
present the deep learning architecture SchNet that is specifically designed to
model atomistic systems by making use of continuous-filter convolutional
layers. We demonstrate the capabilities of SchNet by accurately predicting a
range of properties across chemical space for \emph{molecules and materials}
where our model learns chemically plausible embeddings of atom types across the
periodic table. Finally, we employ SchNet to predict potential-energy surfaces
and energy-conserving force fields for molecular dynamics simulations of small
molecules and perform an exemplary study of the quantum-mechanical properties
of C-fullerene that would have been infeasible with regular ab initio
molecular dynamics
Computational Physics on Graphics Processing Units
The use of graphics processing units for scientific computations is an
emerging strategy that can significantly speed up various different algorithms.
In this review, we discuss advances made in the field of computational physics,
focusing on classical molecular dynamics, and on quantum simulations for
electronic structure calculations using the density functional theory, wave
function techniques, and quantum field theory.Comment: Proceedings of the 11th International Conference, PARA 2012,
Helsinki, Finland, June 10-13, 201
Object orientation and visualization of physics in two dimensions
We present a generalized framework for cellular/lattice based visualizations
in two dimensions based on state of the art computing abstractions. Our
implementation takes the form of a library of reusable functions written in C++
which hides complex graphical programming issues from the user and mimics the
algebraic structure of physics at the Hamiltonian level. Our toolkit is not
just a graphics library but an object analysis of physical systems which
disentangles separate concepts in a faithful analytical way. It could be
rewritten in other languages such as Java and extended to three dimensional
systems straightforwardly. We illustrate the usefulness of our analysis with
implementations of spin-films (the two-dimensional XY model with and without an
external magnetic field) and a model for diffusion through a triangular
lattice.Comment: 12 pages, 10 figure
ASCR/HEP Exascale Requirements Review Report
This draft report summarizes and details the findings, results, and
recommendations derived from the ASCR/HEP Exascale Requirements Review meeting
held in June, 2015. The main conclusions are as follows. 1) Larger, more
capable computing and data facilities are needed to support HEP science goals
in all three frontiers: Energy, Intensity, and Cosmic. The expected scale of
the demand at the 2025 timescale is at least two orders of magnitude -- and in
some cases greater -- than that available currently. 2) The growth rate of data
produced by simulations is overwhelming the current ability, of both facilities
and researchers, to store and analyze it. Additional resources and new
techniques for data analysis are urgently needed. 3) Data rates and volumes
from HEP experimental facilities are also straining the ability to store and
analyze large and complex data volumes. Appropriately configured
leadership-class facilities can play a transformational role in enabling
scientific discovery from these datasets. 4) A close integration of HPC
simulation and data analysis will aid greatly in interpreting results from HEP
experiments. Such an integration will minimize data movement and facilitate
interdependent workflows. 5) Long-range planning between HEP and ASCR will be
required to meet HEP's research needs. To best use ASCR HPC resources the
experimental HEP program needs a) an established long-term plan for access to
ASCR computational and data resources, b) an ability to map workflows onto HPC
resources, c) the ability for ASCR facilities to accommodate workflows run by
collaborations that can have thousands of individual members, d) to transition
codes to the next-generation HPC platforms that will be available at ASCR
facilities, e) to build up and train a workforce capable of developing and
using simulations and analysis to support HEP scientific research on
next-generation systems.Comment: 77 pages, 13 Figures; draft report, subject to further revisio
Atomistic simulations of adiabatic coherent electron transport in triple donor systems
A solid-state analogue of Stimulated Raman Adiabatic Passage can be
implemented in a triple well solid-state system to coherently transport an
electron across the wells with exponentially suppressed occupation in the
central well at any point of time. Termed coherent tunneling adiabatic passage
(CTAP), this method provides a robust way to transfer quantum information
encoded in the electronic spin across a chain of quantum dots or donors. Using
large scale atomistic tight-binding simulations involving over 3.5 million
atoms, we verify the existence of a CTAP pathway in a realistic solid-state
system: gated triple donors in silicon. Realistic gate profiles from commercial
tools were combined with tight-binding methods to simulate gate control of the
donor to donor tunnel barriers in the presence of cross-talk. As CTAP is an
adiabatic protocol, it can be analyzed by solving the time independent problem
at various stages of the pulse - justifying the use of time-independent
tight-binding methods to this problem. Our results show that a three donor CTAP
transfer, with inter-donor spacing of 15 nm can occur on timescales greater
than 23 ps, well within experimentally accessible regimes. The method not only
provides a tool to guide future CTAP experiments, but also illuminates the
possibility of system engineering to enhance control and transfer times.Comment: 8 pages, 5 figure
Grid computing and molecular simulations: the vision of the eMinerals Project
This paper discusses a number of aspects of using grid computing methods in support of molecular simulations, with examples drawn from the eMinerals project. A number of components for a useful grid infrastructure are discussed, including the integration of compute and data grids, automatic metadata capture from simulation studies, interoperability of data between simulation codes, management of data and data accessibility, management of jobs and workflow, and tools to support collaboration. Use of a grid infrastructure also brings certain challenges, which are discussed. These include making use of boundless computing resources, the necessary changes, and the need to be able to manage experimentation
- …