55 research outputs found

    On Routing Disjoint Paths in Bounded Treewidth Graphs

    Get PDF
    We study the problem of routing on disjoint paths in bounded treewidth graphs with both edge and node capacities. The input consists of a capacitated graph GG and a collection of kk source-destination pairs M={(s1,t1),…,(sk,tk)}\mathcal{M} = \{(s_1, t_1), \dots, (s_k, t_k)\}. The goal is to maximize the number of pairs that can be routed subject to the capacities in the graph. A routing of a subset M′\mathcal{M}' of the pairs is a collection P\mathcal{P} of paths such that, for each pair (si,ti)∈M′(s_i, t_i) \in \mathcal{M}', there is a path in P\mathcal{P} connecting sis_i to tit_i. In the Maximum Edge Disjoint Paths (MaxEDP) problem, the graph GG has capacities cap(e)\mathrm{cap}(e) on the edges and a routing P\mathcal{P} is feasible if each edge ee is in at most cap(e)\mathrm{cap}(e) of the paths of P\mathcal{P}. The Maximum Node Disjoint Paths (MaxNDP) problem is the node-capacitated counterpart of MaxEDP. In this paper we obtain an O(r3)O(r^3) approximation for MaxEDP on graphs of treewidth at most rr and a matching approximation for MaxNDP on graphs of pathwidth at most rr. Our results build on and significantly improve the work by Chekuri et al. [ICALP 2013] who obtained an O(r⋅3r)O(r \cdot 3^r) approximation for MaxEDP

    Heuristics for Sparsest Cut Approximations in Network Flow Applications

    Get PDF
    The Maximum Concurrent Flow Problem (MCFP) is a polynomially bounded problem that has been used over the years in a variety of applications. Sometimes it is used to attempt to find the Sparsest Cut, an NP-hard problem, and other times to find communities in Social Network Analysis (SNA) in its hierarchical formulation, the HMCFP. Though it is polynomially bounded, the MCFP quickly grows in space utilization, rendering it useful on only small problems. When it was defined, only a few hundred nodes could be solved, where a few decades later, graphs of one to two thousand nodes can still be too much for modern commodity hardware to handle. This dissertation covers three approaches to heuristics to the MCFP that run significantly faster in practice than the LP formulation with far less memory utilization. The first two approaches are based on the Maximum Adjacency Search (MAS) and apply to both the MCFP and the HMCFP used for community detection. We compare the three approaches to the LP performance in terms of accuracy, runtime, and memory utilization on several classes of synthetic graphs representing potential real-world applications. We find that the heuristics are often correct, and run using orders of magnitude less memory and time

    Deterministic Distributed Algorithms and Lower Bounds in the Hybrid Model

    Get PDF
    The HYBRID model was recently introduced by Augustine et al. [John Augustine et al., 2020] in order to characterize from an algorithmic standpoint the capabilities of networks which combine multiple communication modes. Concretely, it is assumed that the standard LOCAL model of distributed computing is enhanced with the feature of all-to-all communication, but with very limited bandwidth, captured by the node-capacitated clique (NCC). In this work we provide several new insights on the power of hybrid networks for fundamental problems in distributed algorithms. First, we present a deterministic algorithm which solves any problem on a sparse n-node graph in ??(?n) rounds of HYBRID, where the notation ??(?) suppresses polylogarithmic factors of n. We combine this primitive with several sparsification techniques to obtain efficient distributed algorithms for general graphs. Most notably, for the all-pairs shortest paths problem we give deterministic (1 + ?)- and log n/log log n-approximate algorithms for unweighted and weighted graphs respectively with round complexity ??(?n) in HYBRID, closely matching the performance of the state of the art randomized algorithm of Kuhn and Schneider [Kuhn and Schneider, 2020]. Moreover, we make a connection with the Ghaffari-Haeupler framework of low-congestion shortcuts [Mohsen Ghaffari and Bernhard Haeupler, 2016], leading - among others - to a (1 + ?)-approximate algorithm for Min-Cut after ?(polylog (n)) rounds, with high probability, even if we restrict local edges to transfer ?(log n) bits per round. Finally, we prove via a reduction from the set disjointness problem that ??(n^{1/3}) rounds are required to determine the radius of an unweighted graph, as well as a (3/2 - ?)-approximation for weighted graphs. As a byproduct, we show an ??(n) round-complexity lower bound for computing a (4/3 - ?)-approximation of the radius in the broadcast variant of the congested clique, even for unweighted graphs

    Bridge Girth: A Unifying Notion in Network Design

    Full text link
    A classic 1993 paper by Alth\H{o}fer et al. proved a tight reduction from spanners, emulators, and distance oracles to the extremal function γ\gamma of high-girth graphs. This paper initiated a large body of work in network design, in which problems are attacked by reduction to γ\gamma or the analogous extremal function for other girth concepts. In this paper, we introduce and study a new girth concept that we call the bridge girth of path systems, and we show that it can be used to significantly expand and improve this web of connections between girth problems and network design. We prove two kinds of results: 1) We write the maximum possible size of an nn-node, pp-path system with bridge girth >k>k as β(n,p,k)\beta(n, p, k), and we write a certain variant for "ordered" path systems as β∗(n,p,k)\beta^*(n, p, k). We identify several arguments in the literature that implicitly show upper or lower bounds on β,β∗\beta, \beta^*, and we provide some polynomially improvements to these bounds. In particular, we construct a tight lower bound for β(n,p,2)\beta(n, p, 2), and we polynomially improve the upper bounds for β(n,p,4)\beta(n, p, 4) and β∗(n,p,∞)\beta^*(n, p, \infty). 2) We show that many state-of-the-art results in network design can be recovered or improved via black-box reductions to β\beta or β∗\beta^*. Examples include bounds for distance/reachability preservers, exact hopsets, shortcut sets, the flow-cut gaps for directed multicut and sparsest cut, an integrality gap for directed Steiner forest. We believe that the concept of bridge girth can lead to a stronger and more organized map of the research area. Towards this, we leave many open problems, related to both bridge girth reductions and extremal bounds on the size of path systems with high bridge girth

    Vertex sparsification and universal rounding algorithms

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.Cataloged from PDF version of thesis.Includes bibliographical references (p. 125-129).Suppose we are given a gigantic communication network, but are only interested in a small number of nodes (clients). There are many routing problems we could be asked to solve for our clients. Is there a much smaller network - that we could write down on a sheet of paper and put in our pocket - that approximately preserves all the relevant communication properties of the original network? As we will demonstrate, the answer to this question is YES, and we call this smaller network a vertex sparsifier. In fact, if we are asked to solve a sequence of optimization problems characterized by cuts or flows, we can compute a good vertex sparsifier ONCE and discard the original network. We can run our algorithms (or approximation algorithms) on the vertex sparsifier as a proxy - and still recover approximately optimal solutions in the original network. This novel pattern saves both space (because the network we store is much smaller) and time (because our algorithms run on a much smaller graph). Additionally, we apply these ideas to obtain a master theorem for graph partitioning problems - as long as the integrality gap of a standard linear programming relaxation is bounded on trees, then the integrality gap is at most a logarithmic factor larger for general networks. This result implies optimal bounds for many well studied graph partitioning problems as a special case, and even yields optimal bounds for more challenging problems that had not been studied before. Morally, these results are all based on the idea that even though the structure of optimal solutions can be quite complicated, these solution values can be approximated by crude (even linear) functions.by Ankur Moitra.Ph.D

    Vertex sparsification and universal rounding algorithms

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.Cataloged from PDF version of thesis.Includes bibliographical references (p. 125-129).Suppose we are given a gigantic communication network, but are only interested in a small number of nodes (clients). There are many routing problems we could be asked to solve for our clients. Is there a much smaller network - that we could write down on a sheet of paper and put in our pocket - that approximately preserves all the relevant communication properties of the original network? As we will demonstrate, the answer to this question is YES, and we call this smaller network a vertex sparsifier. In fact, if we are asked to solve a sequence of optimization problems characterized by cuts or flows, we can compute a good vertex sparsifier ONCE and discard the original network. We can run our algorithms (or approximation algorithms) on the vertex sparsifier as a proxy - and still recover approximately optimal solutions in the original network. This novel pattern saves both space (because the network we store is much smaller) and time (because our algorithms run on a much smaller graph). Additionally, we apply these ideas to obtain a master theorem for graph partitioning problems - as long as the integrality gap of a standard linear programming relaxation is bounded on trees, then the integrality gap is at most a logarithmic factor larger for general networks. This result implies optimal bounds for many well studied graph partitioning problems as a special case, and even yields optimal bounds for more challenging problems that had not been studied before. Morally, these results are all based on the idea that even though the structure of optimal solutions can be quite complicated, these solution values can be approximated by crude (even linear) functions.by Ankur Moitra.Ph.D
    • …
    corecore