6,927 research outputs found

    Fast and Deterministic Approximations for k-Cut

    Get PDF
    In an undirected graph, a k-cut is a set of edges whose removal breaks the graph into at least k connected components. The minimum weight k-cut can be computed in n^O(k) time, but when k is treated as part of the input, computing the minimum weight k-cut is NP-Hard [Goldschmidt and Hochbaum, 1994]. For poly(m,n,k)-time algorithms, the best possible approximation factor is essentially 2 under the small set expansion hypothesis [Manurangsi, 2017]. Saran and Vazirani [1995] showed that a (2 - 2/k)-approximately minimum weight k-cut can be computed via O(k) minimum cuts, which implies a O~(km) randomized running time via the nearly linear time randomized min-cut algorithm of Karger [2000]. Nagamochi and Kamidoi [2007] showed that a (2 - 2/k)-approximately minimum weight k-cut can be computed deterministically in O(mn + n^2 log n) time. These results prompt two basic questions. The first concerns the role of randomization. Is there a deterministic algorithm for 2-approximate k-cuts matching the randomized running time of O~(km)? The second question qualitatively compares minimum cut to 2-approximate minimum k-cut. Can 2-approximate k-cuts be computed as fast as the minimum cut - in O~(m) randomized time? We give a deterministic approximation algorithm that computes (2 + eps)-minimum k-cuts in O(m log^3 n / eps^2) time, via a (1 + eps)-approximation for an LP relaxation of k-cut

    Approximating the Held-Karp Bound for Metric TSP in Nearly Linear Time

    Full text link
    We give a nearly linear time randomized approximation scheme for the Held-Karp bound [Held and Karp, 1970] for metric TSP. Formally, given an undirected edge-weighted graph GG on mm edges and ϵ>0\epsilon > 0, the algorithm outputs in O(mlog4n/ϵ2)O(m \log^4n /\epsilon^2) time, with high probability, a (1+ϵ)(1+\epsilon)-approximation to the Held-Karp bound on the metric TSP instance induced by the shortest path metric on GG. The algorithm can also be used to output a corresponding solution to the Subtour Elimination LP. We substantially improve upon the O(m2log2(m)/ϵ2)O(m^2 \log^2(m)/\epsilon^2) running time achieved previously by Garg and Khandekar. The LP solution can be used to obtain a fast randomized (32+ϵ)\big(\frac{3}{2} + \epsilon\big)-approximation for metric TSP which improves upon the running time of previous implementations of Christofides' algorithm

    Eight-Fifth Approximation for TSP Paths

    Full text link
    We prove the approximation ratio 8/5 for the metric {s,t}\{s,t\}-path-TSP problem, and more generally for shortest connected TT-joins. The algorithm that achieves this ratio is the simple "Best of Many" version of Christofides' algorithm (1976), suggested by An, Kleinberg and Shmoys (2012), which consists in determining the best Christofides {s,t}\{s,t\}-tour out of those constructed from a family \Fscr_{>0} of trees having a convex combination dominated by an optimal solution xx^* of the fractional relaxation. They give the approximation guarantee 5+12\frac{\sqrt{5}+1}{2} for such an {s,t}\{s,t\}-tour, which is the first improvement after the 5/3 guarantee of Hoogeveen's Christofides type algorithm (1991). Cheriyan, Friggstad and Gao (2012) extended this result to a 13/8-approximation of shortest connected TT-joins, for T4|T|\ge 4. The ratio 8/5 is proved by simplifying and improving the approach of An, Kleinberg and Shmoys that consists in completing x/2x^*/2 in order to dominate the cost of "parity correction" for spanning trees. We partition the edge-set of each spanning tree in \Fscr_{>0} into an {s,t}\{s,t\}-path (or more generally, into a TT-join) and its complement, which induces a decomposition of xx^*. This decomposition can be refined and then efficiently used to complete x/2x^*/2 without using linear programming or particular properties of TT, but by adding to each cut deficient for x/2x^*/2 an individually tailored explicitly given vector, inherent in xx^*. A simple example shows that the Best of Many Christofides algorithm may not find a shorter {s,t}\{s,t\}-tour than 3/2 times the incidentally common optima of the problem and of its fractional relaxation.Comment: 15 pages, corrected typos in citations, minor change

    Almost-Tight Distributed Minimum Cut Algorithms

    Full text link
    We study the problem of computing the minimum cut in a weighted distributed message-passing networks (the CONGEST model). Let λ\lambda be the minimum cut, nn be the number of nodes in the network, and DD be the network diameter. Our algorithm can compute λ\lambda exactly in O((nlogn+D)λ4log2n)O((\sqrt{n} \log^{*} n+D)\lambda^4 \log^2 n) time. To the best of our knowledge, this is the first paper that explicitly studies computing the exact minimum cut in the distributed setting. Previously, non-trivial sublinear time algorithms for this problem are known only for unweighted graphs when λ3\lambda\leq 3 due to Pritchard and Thurimella's O(D)O(D)-time and O(D+n1/2logn)O(D+n^{1/2}\log^* n)-time algorithms for computing 22-edge-connected and 33-edge-connected components. By using the edge sampling technique of Karger's, we can convert this algorithm into a (1+ϵ)(1+\epsilon)-approximation O((nlogn+D)ϵ5log3n)O((\sqrt{n}\log^{*} n+D)\epsilon^{-5}\log^3 n)-time algorithm for any ϵ>0\epsilon>0. This improves over the previous (2+ϵ)(2+\epsilon)-approximation O((nlogn+D)ϵ5log2nloglogn)O((\sqrt{n}\log^{*} n+D)\epsilon^{-5}\log^2 n\log\log n)-time algorithm and O(ϵ1)O(\epsilon^{-1})-approximation O(D+n12+ϵpolylogn)O(D+n^{\frac{1}{2}+\epsilon} \mathrm{poly}\log n)-time algorithm of Ghaffari and Kuhn. Due to the lower bound of Ω(D+n1/2/logn)\Omega(D+n^{1/2}/\log n) by Das Sarma et al. which holds for any approximation algorithm, this running time is tight up to a polylogn \mathrm{poly}\log n factor. To get the stated running time, we developed an approximation algorithm which combines the ideas of Thorup's algorithm and Matula's contraction algorithm. It saves an ϵ9log7n\epsilon^{-9}\log^{7} n factor as compared to applying Thorup's tree packing theorem directly. Then, we combine Kutten and Peleg's tree partitioning algorithm and Karger's dynamic programming to achieve an efficient distributed algorithm that finds the minimum cut when we are given a spanning tree that crosses the minimum cut exactly once

    Graph Sparsification by Edge-Connectivity and Random Spanning Trees

    Full text link
    We present new approaches to constructing graph sparsifiers --- weighted subgraphs for which every cut has the same value as the original graph, up to a factor of (1±ϵ)(1 \pm \epsilon). Our first approach independently samples each edge uvuv with probability inversely proportional to the edge-connectivity between uu and vv. The fact that this approach produces a sparsifier resolves a question posed by Bencz\'ur and Karger (2002). Concurrent work of Hariharan and Panigrahi also resolves this question. Our second approach constructs a sparsifier by forming the union of several uniformly random spanning trees. Both of our approaches produce sparsifiers with O(nlog2(n)/ϵ2)O(n \log^2(n)/\epsilon^2) edges. Our proofs are based on extensions of Karger's contraction algorithm, which may be of independent interest

    Finite Volume Spaces and Sparsification

    Full text link
    We introduce and study finite dd-volumes - the high dimensional generalization of finite metric spaces. Having developed a suitable combinatorial machinery, we define 1\ell_1-volumes and show that they contain Euclidean volumes and hypertree volumes. We show that they can approximate any dd-volume with O(nd)O(n^d) multiplicative distortion. On the other hand, contrary to Bourgain's theorem for d=1d=1, there exists a 22-volume that on nn vertices that cannot be approximated by any 1\ell_1-volume with distortion smaller than Ω~(n1/5)\tilde{\Omega}(n^{1/5}). We further address the problem of 1\ell_1-dimension reduction in the context of 1\ell_1 volumes, and show that this phenomenon does occur, although not to the same striking degree as it does for Euclidean metrics and volumes. In particular, we show that any 1\ell_1 metric on nn points can be (1+ϵ)(1+ \epsilon)-approximated by a sum of O(n/ϵ2)O(n/\epsilon^2) cut metrics, improving over the best previously known bound of O(nlogn)O(n \log n) due to Schechtman. In order to deal with dimension reduction, we extend the techniques and ideas introduced by Karger and Bencz{\'u}r, and Spielman et al.~in the context of graph Sparsification, and develop general methods with a wide range of applications.Comment: previous revision was the wrong file: the new revision: changed (extended considerably) the treatment of finite volumes (see revised abstract). Inserted new applications for the sparsification technique
    corecore