211,418 research outputs found
Theoretically Efficient Parallel Graph Algorithms Can Be Fast and Scalable
There has been significant recent interest in parallel graph processing due
to the need to quickly analyze the large graphs available today. Many graph
codes have been designed for distributed memory or external memory. However,
today even the largest publicly-available real-world graph (the Hyperlink Web
graph with over 3.5 billion vertices and 128 billion edges) can fit in the
memory of a single commodity multicore server. Nevertheless, most experimental
work in the literature report results on much smaller graphs, and the ones for
the Hyperlink graph use distributed or external memory. Therefore, it is
natural to ask whether we can efficiently solve a broad class of graph problems
on this graph in memory.
This paper shows that theoretically-efficient parallel graph algorithms can
scale to the largest publicly-available graphs using a single machine with a
terabyte of RAM, processing them in minutes. We give implementations of
theoretically-efficient parallel algorithms for 20 important graph problems. We
also present the optimizations and techniques that we used in our
implementations, which were crucial in enabling us to process these large
graphs quickly. We show that the running times of our implementations
outperform existing state-of-the-art implementations on the largest real-world
graphs. For many of the problems that we consider, this is the first time they
have been solved on graphs at this scale. We have made the implementations
developed in this work publicly-available as the Graph-Based Benchmark Suite
(GBBS).Comment: This is the full version of the paper appearing in the ACM Symposium
on Parallelism in Algorithms and Architectures (SPAA), 201
Portfolio-based Planning: State of the Art, Common Practice and Open Challenges
In recent years the field of automated planning has significantly
advanced and several powerful domain-independent
planners have been developed. However, none of these systems
clearly outperforms all the others in every known
benchmark domain. This observation motivated the idea of
configuring and exploiting a portfolio of planners to perform
better than any individual planner: some recent planning systems
based on this idea achieved significantly good results in
experimental analysis and International Planning Competitions.
Such results let us suppose that future challenges of the
Automated Planning community will converge on designing
different approaches for combining existing planning algorithms.
This paper reviews existing techniques and provides an exhaustive
guide to portfolio-based planning. In addition, the
paper outlines open issues of existing approaches and highlights
possible future evolution of these techniques
- …