5,404 research outputs found
Cluster Before You Hallucinate: Approximating Node-Capacitated Network Design and Energy Efficient Routing
We consider circuit routing with an objective of minimizing energy, in a
network of routers that are speed scalable and that may be shutdown when idle.
We consider both multicast routing and unicast routing. It is known that this
energy minimization problem can be reduced to a capacitated flow network design
problem, where vertices have a common capacity but arbitrary costs, and the
goal is to choose a minimum cost collection of vertices whose induced subgraph
will support the specified flow requirements. For the multicast (single-sink)
capacitated design problem we give a polynomial-time algorithm that is
O(log^3n)-approximate with O(log^4 n) congestion. This translates back to a
O(log ^(4{\alpha}+3) n)-approximation for the multicast energy-minimization
routing problem, where {\alpha} is the polynomial exponent in the dynamic power
used by a router. For the unicast (multicommodity) capacitated design problem
we give a polynomial-time algorithm that is O(log^5 n)-approximate with
O(log^12 n) congestion, which translates back to a O(log^(12{\alpha}+5)
n)-approximation for the unicast energy-minimization routing problem.Comment: 22 pages (full version of STOC 2014 paper
-Approximation Algorithm for Directed Steiner Tree: A Tight Quasi-Polynomial-Time Algorithm
In the Directed Steiner Tree (DST) problem we are given an -vertex
directed edge-weighted graph, a root , and a collection of terminal
nodes. Our goal is to find a minimum-cost arborescence that contains a directed
path from to every terminal. We present an -approximation algorithm for DST that runs in
quasi-polynomial-time. By adjusting the parameters in the hardness result of
Halperin and Krauthgamer, we show the matching lower bound of
for the class of quasi-polynomial-time
algorithms. This is the first improvement on the DST problem since the
classical quasi-polynomial-time approximation algorithm by
Charikar et al. (The paper erroneously claims an approximation due
to a mistake in prior work.)
Our approach is based on two main ingredients. First, we derive an
approximation preserving reduction to the Label-Consistent Subtree (LCST)
problem. The LCST instance has quasi-polynomial size and logarithmic height. We
remark that, in contrast, Zelikovsky's heigh-reduction theorem used in all
prior work on DST achieves a reduction to a tree instance of the related Group
Steiner Tree (GST) problem of similar height, however losing a logarithmic
factor in the approximation ratio. Our second ingredient is an LP-rounding
algorithm to approximately solve LCST instances, which is inspired by the
framework developed by Rothvo{\ss}. We consider a Sherali-Adams lifting of a
proper LP relaxation of LCST. Our rounding algorithm proceeds level by level
from the root to the leaves, rounding and conditioning each time on a proper
subset of label variables. A small enough (namely, polylogarithmic) number of
Sherali-Adams lifting levels is sufficient to condition up to the leaves
Integrality Gap of the Hypergraphic Relaxation of Steiner Trees: a short proof of a 1.55 upper bound
Recently Byrka, Grandoni, Rothvoss and Sanita (at STOC 2010) gave a
1.39-approximation for the Steiner tree problem, using a hypergraph-based
linear programming relaxation. They also upper-bounded its integrality gap by
1.55. We describe a shorter proof of the same integrality gap bound, by
applying some of their techniques to a randomized loss-contracting algorithm
Greedy Algorithms for Steiner Forest
In the Steiner Forest problem, we are given terminal pairs ,
and need to find the cheapest subgraph which connects each of the terminal
pairs together. In 1991, Agrawal, Klein, and Ravi, and Goemans and Williamson
gave primal-dual constant-factor approximation algorithms for this problem;
until now, the only constant-factor approximations we know are via linear
programming relaxations.
We consider the following greedy algorithm: Given terminal pairs in a metric
space, call a terminal "active" if its distance to its partner is non-zero.
Pick the two closest active terminals (say ), set the distance
between them to zero, and buy a path connecting them. Recompute the metric, and
repeat. Our main result is that this algorithm is a constant-factor
approximation.
We also use this algorithm to give new, simpler constructions of cost-sharing
schemes for Steiner forest. In particular, the first "group-strict" cost-shares
for this problem implies a very simple combinatorial sampling-based algorithm
for stochastic Steiner forest
Two-Level Rectilinear Steiner Trees
Given a set of terminals in the plane and a partition of into
subsets , a two-level rectilinear Steiner tree consists of a
rectilinear Steiner tree connecting the terminals in each set
() and a top-level tree connecting the trees . The goal is to minimize the total length of all trees. This problem
arises naturally in the design of low-power physical implementations of parity
functions on a computer chip.
For bounded we present a polynomial time approximation scheme (PTAS) that
is based on Arora's PTAS for rectilinear Steiner trees after lifting each
partition into an extra dimension. For the general case we propose an algorithm
that predetermines a connection point for each and
().
Then, we apply any approximation algorithm for minimum rectilinear Steiner
trees in the plane to compute each and independently.
This gives us a -factor approximation with a running time of
suitable for fast practical computations. The
approximation factor reduces to by applying Arora's approximation scheme
in the plane
Squarepants in a Tree: Sum of Subtree Clustering and Hyperbolic Pants Decomposition
We provide efficient constant factor approximation algorithms for the
problems of finding a hierarchical clustering of a point set in any metric
space, minimizing the sum of minimimum spanning tree lengths within each
cluster, and in the hyperbolic or Euclidean planes, minimizing the sum of
cluster perimeters. Our algorithms for the hyperbolic and Euclidean planes can
also be used to provide a pants decomposition, that is, a set of disjoint
simple closed curves partitioning the plane minus the input points into subsets
with exactly three boundary components, with approximately minimum total
length. In the Euclidean case, these curves are squares; in the hyperbolic
case, they combine our Euclidean square pants decomposition with our tree
clustering method for general metric spaces.Comment: 22 pages, 14 figures. This version replaces the proof of what is now
Lemma 5.2, as the previous proof was erroneou
- …