43,326 research outputs found
Pregelix: Big(ger) Graph Analytics on A Dataflow Engine
There is a growing need for distributed graph processing systems that are
capable of gracefully scaling to very large graph datasets. Unfortunately, this
challenge has not been easily met due to the intense memory pressure imposed by
process-centric, message passing designs that many graph processing systems
follow. Pregelix is a new open source distributed graph processing system that
is based on an iterative dataflow design that is better tuned to handle both
in-memory and out-of-core workloads. As such, Pregelix offers improved
performance characteristics and scaling properties over current open source
systems (e.g., we have seen up to 15x speedup compared to Apache Giraph and up
to 35x speedup compared to distributed GraphLab), and makes more effective use
of available machine resources to support Big(ger) Graph Analytics
Quantum Monte Carlo for large chemical systems: Implementing efficient strategies for petascale platforms and beyond
Various strategies to implement efficiently QMC simulations for large
chemical systems are presented. These include: i.) the introduction of an
efficient algorithm to calculate the computationally expensive Slater matrices.
This novel scheme is based on the use of the highly localized character of
atomic Gaussian basis functions (not the molecular orbitals as usually done),
ii.) the possibility of keeping the memory footprint minimal, iii.) the
important enhancement of single-core performance when efficient optimization
tools are employed, and iv.) the definition of a universal, dynamic,
fault-tolerant, and load-balanced computational framework adapted to all kinds
of computational platforms (massively parallel machines, clusters, or
distributed grids). These strategies have been implemented in the QMC=Chem code
developed at Toulouse and illustrated with numerical applications on small
peptides of increasing sizes (158, 434, 1056 and 1731 electrons). Using 10k-80k
computing cores of the Curie machine (GENCI-TGCC-CEA, France) QMC=Chem has been
shown to be capable of running at the petascale level, thus demonstrating that
for this machine a large part of the peak performance can be achieved.
Implementation of large-scale QMC simulations for future exascale platforms
with a comparable level of efficiency is expected to be feasible
A model and framework for reliable build systems
Reliable and fast builds are essential for rapid turnaround during
development and testing. Popular existing build systems rely on correct manual
specification of build dependencies, which can lead to invalid build outputs
and nondeterminism. We outline the challenges of developing reliable build
systems and explore the design space for their implementation, with a focus on
non-distributed, incremental, parallel build systems. We define a general model
for resources accessed by build tasks and show its correspondence to the
implementation technique of minimum information libraries, APIs that return no
information that the application doesn't plan to use. We also summarize
preliminary experimental results from several prototype build managers
Shortest, Fastest, and Foremost Broadcast in Dynamic Networks
Highly dynamic networks rarely offer end-to-end connectivity at a given time.
Yet, connectivity in these networks can be established over time and space,
based on temporal analogues of multi-hop paths (also called {\em journeys}).
Attempting to optimize the selection of the journeys in these networks
naturally leads to the study of three cases: shortest (minimum hop), fastest
(minimum duration), and foremost (earliest arrival) journeys. Efficient
centralized algorithms exists to compute all cases, when the full knowledge of
the network evolution is given.
In this paper, we study the {\em distributed} counterparts of these problems,
i.e. shortest, fastest, and foremost broadcast with termination detection
(TDB), with minimal knowledge on the topology.
We show that the feasibility of each of these problems requires distinct
features on the evolution, through identifying three classes of dynamic graphs
wherein the problems become gradually feasible: graphs in which the
re-appearance of edges is {\em recurrent} (class R), {\em bounded-recurrent}
(B), or {\em periodic} (P), together with specific knowledge that are
respectively (the number of nodes), (a bound on the recurrence
time), and (the period). In these classes it is not required that all pairs
of nodes get in contact -- only that the overall {\em footprint} of the graph
is connected over time.
Our results, together with the strict inclusion between , , and ,
implies a feasibility order among the three variants of the problem, i.e.
TDB[foremost] requires weaker assumptions on the topology dynamics than
TDB[shortest], which itself requires less than TDB[fastest]. Reversely, these
differences in feasibility imply that the computational powers of ,
, and also form a strict hierarchy
- …