58,734 research outputs found
Theoretically Efficient Parallel Graph Algorithms Can Be Fast and Scalable
There has been significant recent interest in parallel graph processing due
to the need to quickly analyze the large graphs available today. Many graph
codes have been designed for distributed memory or external memory. However,
today even the largest publicly-available real-world graph (the Hyperlink Web
graph with over 3.5 billion vertices and 128 billion edges) can fit in the
memory of a single commodity multicore server. Nevertheless, most experimental
work in the literature report results on much smaller graphs, and the ones for
the Hyperlink graph use distributed or external memory. Therefore, it is
natural to ask whether we can efficiently solve a broad class of graph problems
on this graph in memory.
This paper shows that theoretically-efficient parallel graph algorithms can
scale to the largest publicly-available graphs using a single machine with a
terabyte of RAM, processing them in minutes. We give implementations of
theoretically-efficient parallel algorithms for 20 important graph problems. We
also present the optimizations and techniques that we used in our
implementations, which were crucial in enabling us to process these large
graphs quickly. We show that the running times of our implementations
outperform existing state-of-the-art implementations on the largest real-world
graphs. For many of the problems that we consider, this is the first time they
have been solved on graphs at this scale. We have made the implementations
developed in this work publicly-available as the Graph-Based Benchmark Suite
(GBBS).Comment: This is the full version of the paper appearing in the ACM Symposium
on Parallelism in Algorithms and Architectures (SPAA), 201
Fine-grained dichotomies for the Tutte plane and Boolean #CSP
Jaeger, Vertigan, and Welsh [15] proved a dichotomy for the complexity of
evaluating the Tutte polynomial at fixed points: The evaluation is #P-hard
almost everywhere, and the remaining points admit polynomial-time algorithms.
Dell, Husfeldt, and Wahl\'en [9] and Husfeldt and Taslaman [12], in combination
with Curticapean [7], extended the #P-hardness results to tight lower bounds
under the counting exponential time hypothesis #ETH, with the exception of the
line , which was left open. We complete the dichotomy theorem for the
Tutte polynomial under #ETH by proving that the number of all acyclic subgraphs
of a given -vertex graph cannot be determined in time unless
#ETH fails.
Another dichotomy theorem we strengthen is the one of Creignou and Hermann
[6] for counting the number of satisfying assignments to a constraint
satisfaction problem instance over the Boolean domain. We prove that all
#P-hard cases are also hard under #ETH. The main ingredient is to prove that
the number of independent sets in bipartite graphs with vertices cannot be
computed in time unless #ETH fails. In order to prove our results,
we use the block interpolation idea by Curticapean [7] and transfer it to
systems of linear equations that might not directly correspond to
interpolation.Comment: 16 pages, 1 figur
Large induced subgraphs via triangulations and CMSO
We obtain an algorithmic meta-theorem for the following optimization problem.
Let \phi\ be a Counting Monadic Second Order Logic (CMSO) formula and t be an
integer. For a given graph G, the task is to maximize |X| subject to the
following: there is a set of vertices F of G, containing X, such that the
subgraph G[F] induced by F is of treewidth at most t, and structure (G[F],X)
models \phi.
Some special cases of this optimization problem are the following generic
examples. Each of these cases contains various problems as a special subcase:
1) "Maximum induced subgraph with at most l copies of cycles of length 0
modulo m", where for fixed nonnegative integers m and l, the task is to find a
maximum induced subgraph of a given graph with at most l vertex-disjoint cycles
of length 0 modulo m.
2) "Minimum \Gamma-deletion", where for a fixed finite set of graphs \Gamma\
containing a planar graph, the task is to find a maximum induced subgraph of a
given graph containing no graph from \Gamma\ as a minor.
3) "Independent \Pi-packing", where for a fixed finite set of connected
graphs \Pi, the task is to find an induced subgraph G[F] of a given graph G
with the maximum number of connected components, such that each connected
component of G[F] is isomorphic to some graph from \Pi.
We give an algorithm solving the optimization problem on an n-vertex graph G
in time O(#pmc n^{t+4} f(t,\phi)), where #pmc is the number of all potential
maximal cliques in G and f is a function depending of t and \phi\ only. We also
show how a similar running time can be obtained for the weighted version of the
problem. Pipelined with known bounds on the number of potential maximal
cliques, we deduce that our optimization problem can be solved in time
O(1.7347^n) for arbitrary graphs, and in polynomial time for graph classes with
polynomial number of minimal separators
Exponential Time Complexity of Weighted Counting of Independent Sets
We consider weighted counting of independent sets using a rational weight x:
Given a graph with n vertices, count its independent sets such that each set of
size k contributes x^k. This is equivalent to computation of the partition
function of the lattice gas with hard-core self-repulsion and hard-core pair
interaction. We show the following conditional lower bounds: If counting the
satisfying assignments of a 3-CNF formula in n variables (#3SAT) needs time
2^{\Omega(n)} (i.e. there is a c>0 such that no algorithm can solve #3SAT in
time 2^{cn}), counting the independent sets of size n/3 of an n-vertex graph
needs time 2^{\Omega(n)} and weighted counting of independent sets needs time
2^{\Omega(n/log^3 n)} for all rational weights x\neq 0.
We have two technical ingredients: The first is a reduction from 3SAT to
independent sets that preserves the number of solutions and increases the
instance size only by a constant factor. Second, we devise a combination of
vertex cloning and path addition. This graph transformation allows us to adapt
a recent technique by Dell, Husfeldt, and Wahlen which enables interpolation by
a family of reductions, each of which increases the instance size only
polylogarithmically.Comment: Introduction revised, differences between versions of counting
independent sets stated more precisely, minor improvements. 14 page
FPTAS for Hardcore and Ising Models on Hypergraphs
Hardcore and Ising models are two most important families of two state spin
systems in statistic physics. Partition function of spin systems is the center
concept in statistic physics which connects microscopic particles and their
interactions with their macroscopic and statistical properties of materials
such as energy, entropy, ferromagnetism, etc. If each local interaction of the
system involves only two particles, the system can be described by a graph. In
this case, fully polynomial-time approximation scheme (FPTAS) for computing the
partition function of both hardcore and anti-ferromagnetic Ising model was
designed up to the uniqueness condition of the system. These result are the
best possible since approximately computing the partition function beyond this
threshold is NP-hard. In this paper, we generalize these results to general
physics systems, where each local interaction may involves multiple particles.
Such systems are described by hypergraphs. For hardcore model, we also provide
FPTAS up to the uniqueness condition, and for anti-ferromagnetic Ising model,
we obtain FPTAS where a slightly stronger condition holds
Fast Algorithms for Parameterized Problems with Relaxed Disjointness Constraints
In parameterized complexity, it is a natural idea to consider different
generalizations of classic problems. Usually, such generalization are obtained
by introducing a "relaxation" variable, where the original problem corresponds
to setting this variable to a constant value. For instance, the problem of
packing sets of size at most into a given universe generalizes the Maximum
Matching problem, which is recovered by taking . Most often, the
complexity of the problem increases with the relaxation variable, but very
recently Abasi et al. have given a surprising example of a problem ---
-Simple -Path --- that can be solved by a randomized algorithm with
running time . That is, the complexity of the
problem decreases with . In this paper we pursue further the direction
sketched by Abasi et al. Our main contribution is a derandomization tool that
provides a deterministic counterpart of the main technical result of Abasi et
al.: the algorithm for -Monomial
Detection, which is the problem of finding a monomial of total degree and
individual degrees at most in a polynomial given as an arithmetic circuit.
Our technique works for a large class of circuits, and in particular it can be
used to derandomize the result of Abasi et al. for -Simple -Path. On our
way to this result we introduce the notion of representative sets for
multisets, which may be of independent interest. Finally, we give two more
examples of problems that were already studied in the literature, where the
same relaxation phenomenon happens. The first one is a natural relaxation of
the Set Packing problem, where we allow the packed sets to overlap at each
element at most times. The second one is Degree Bounded Spanning Tree,
where we seek for a spanning tree of the graph with a small maximum degree
Estimating the Spectrum in Computed Tomography Via Kullback–Leibler Divergence Constrained Optimization
Purpose
We study the problem of spectrum estimation from transmission data of a known phantom. The goal is to reconstruct an x‐ray spectrum that can accurately model the x‐ray transmission curves and reflects a realistic shape of the typical energy spectra of the CT system. Methods
Spectrum estimation is posed as an optimization problem with x‐ray spectrum as unknown variables, and a Kullback–Leibler (KL)‐divergence constraint is employed to incorporate prior knowledge of the spectrum and enhance numerical stability of the estimation process. The formulated constrained optimization problem is convex and can be solved efficiently by use of the exponentiated‐gradient (EG) algorithm. We demonstrate the effectiveness of the proposed approach on the simulated and experimental data. The comparison to the expectation–maximization (EM) method is also discussed. Results
In simulations, the proposed algorithm is seen to yield x‐ray spectra that closely match the ground truth and represent the attenuation process of x‐ray photons in materials, both included and not included in the estimation process. In experiments, the calculated transmission curve is in good agreement with the measured transmission curve, and the estimated spectra exhibits physically realistic looking shapes. The results further show the comparable performance between the proposed optimization‐based approach and EM. Conclusions
Our formulation of a constrained optimization provides an interpretable and flexible framework for spectrum estimation. Moreover, a KL‐divergence constraint can include a prior spectrum and appears to capture important features of x‐ray spectrum, allowing accurate and robust estimation of x‐ray spectrum in CT imaging
- …