5,403 research outputs found
Parameterized Approximation Schemes for Steiner Trees with Small Number of Steiner Vertices
We study the Steiner Tree problem, in which a set of terminal vertices needs
to be connected in the cheapest possible way in an edge-weighted graph. This
problem has been extensively studied from the viewpoint of approximation and
also parametrization. In particular, on one hand Steiner Tree is known to be
APX-hard, and W[2]-hard on the other, if parameterized by the number of
non-terminals (Steiner vertices) in the optimum solution. In contrast to this
we give an efficient parameterized approximation scheme (EPAS), which
circumvents both hardness results. Moreover, our methods imply the existence of
a polynomial size approximate kernelization scheme (PSAKS) for the considered
parameter.
We further study the parameterized approximability of other variants of
Steiner Tree, such as Directed Steiner Tree and Steiner Forest. For neither of
these an EPAS is likely to exist for the studied parameter: for Steiner Forest
an easy observation shows that the problem is APX-hard, even if the input graph
contains no Steiner vertices. For Directed Steiner Tree we prove that
approximating within any function of the studied parameter is W[1]-hard.
Nevertheless, we show that an EPAS exists for Unweighted Directed Steiner Tree,
but a PSAKS does not. We also prove that there is an EPAS and a PSAKS for
Steiner Forest if in addition to the number of Steiner vertices, the number of
connected components of an optimal solution is considered to be a parameter.Comment: 23 pages, 6 figures An extended abstract appeared in proceedings of
STACS 201
Low Diameter Graph Decompositions by Approximate Distance Computation
In many models for large-scale computation, decomposition of the problem is key to efficient algorithms. For distance-related graph problems, it is often crucial that such a decomposition results in clusters of small diameter, while the probability that an edge is cut by the decomposition scales linearly with the length of the edge. There is a large body of literature on low diameter graph decomposition with small edge cutting probabilities, with all existing techniques heavily building on single source shortest paths (SSSP) computations. Unfortunately, in many theoretical models for large-scale computations, the SSSP task constitutes a complexity bottleneck. Therefore, it is desirable to replace exact SSSP computations with approximate ones. However this imposes a fundamental challenge since the existing constructions of low diameter graph decomposition with small edge cutting probabilities inherently rely on the subtractive form of the triangle inequality, which fails to hold under distance approximation.
The current paper overcomes this obstacle by developing a technique termed blurry ball growing. By combining this technique with a clever algorithmic idea of Miller et al. (SPAA 2013), we obtain a construction of low diameter decompositions with small edge cutting probabilities which replaces exact SSSP computations by (a small number of) approximate ones. The utility of our approach is showcased by deriving efficient algorithms that work in the CONGEST, PRAM, and semi-streaming models of computation. As an application, we obtain metric tree embedding algorithms in the vein of Bartal (FOCS 1996) whose computational complexities in these models are optimal up to polylogarithmic factors. Our embeddings have the additional useful property that the tree can be mapped back to the original graph such that each edge is "used" only logaritmically many times, which is of interest for capacitated problems and simulating CONGEST algorithms on the tree into which the graph is embedded
Almost-Tight Distributed Minimum Cut Algorithms
We study the problem of computing the minimum cut in a weighted distributed
message-passing networks (the CONGEST model). Let be the minimum cut,
be the number of nodes in the network, and be the network diameter. Our
algorithm can compute exactly in time. To the best of our knowledge, this is the first paper that
explicitly studies computing the exact minimum cut in the distributed setting.
Previously, non-trivial sublinear time algorithms for this problem are known
only for unweighted graphs when due to Pritchard and
Thurimella's -time and -time algorithms for
computing -edge-connected and -edge-connected components.
By using the edge sampling technique of Karger's, we can convert this
algorithm into a -approximation -time algorithm for any . This improves
over the previous -approximation -time algorithm and
-approximation -time algorithm of Ghaffari and Kuhn. Due to the lower
bound of by Das Sarma et al. which holds for any
approximation algorithm, this running time is tight up to a factor.
To get the stated running time, we developed an approximation algorithm which
combines the ideas of Thorup's algorithm and Matula's contraction algorithm. It
saves an factor as compared to applying Thorup's tree
packing theorem directly. Then, we combine Kutten and Peleg's tree partitioning
algorithm and Karger's dynamic programming to achieve an efficient distributed
algorithm that finds the minimum cut when we are given a spanning tree that
crosses the minimum cut exactly once
The Minimum Wiener Connector
The Wiener index of a graph is the sum of all pairwise shortest-path
distances between its vertices. In this paper we study the novel problem of
finding a minimum Wiener connector: given a connected graph and a set
of query vertices, find a subgraph of that connects all
query vertices and has minimum Wiener index.
We show that The Minimum Wiener Connector admits a polynomial-time (albeit
impractical) exact algorithm for the special case where the number of query
vertices is bounded. We show that in general the problem is NP-hard, and has no
PTAS unless . Our main contribution is a
constant-factor approximation algorithm running in time
.
A thorough experimentation on a large variety of real-world graphs confirms
that our method returns smaller and denser solutions than other methods, and
does so by adding to the query set a small number of important vertices
(i.e., vertices with high centrality).Comment: Published in Proceedings of the 2015 ACM SIGMOD International
Conference on Management of Dat
Cluster Before You Hallucinate: Approximating Node-Capacitated Network Design and Energy Efficient Routing
We consider circuit routing with an objective of minimizing energy, in a
network of routers that are speed scalable and that may be shutdown when idle.
We consider both multicast routing and unicast routing. It is known that this
energy minimization problem can be reduced to a capacitated flow network design
problem, where vertices have a common capacity but arbitrary costs, and the
goal is to choose a minimum cost collection of vertices whose induced subgraph
will support the specified flow requirements. For the multicast (single-sink)
capacitated design problem we give a polynomial-time algorithm that is
O(log^3n)-approximate with O(log^4 n) congestion. This translates back to a
O(log ^(4{\alpha}+3) n)-approximation for the multicast energy-minimization
routing problem, where {\alpha} is the polynomial exponent in the dynamic power
used by a router. For the unicast (multicommodity) capacitated design problem
we give a polynomial-time algorithm that is O(log^5 n)-approximate with
O(log^12 n) congestion, which translates back to a O(log^(12{\alpha}+5)
n)-approximation for the unicast energy-minimization routing problem.Comment: 22 pages (full version of STOC 2014 paper
- …