125,186 research outputs found

    Distributed Approximation on Power Graphs

    Full text link
    We investigate graph problems in the following setting: we are given a graph GG and we are required to solve a problem on G2G^2. While we focus mostly on exploring this theme in the distributed CONGEST model, we show new results and surprising connections to the centralized model of computation. In the CONGEST model, it is natural to expect that problems on G2G^2 would be quite difficult to solve efficiently on GG, due to congestion. However, we show that the picture is both more complicated and more interesting. Specifically, we encounter two phenomena acting in opposing directions: (i) slowdown due to congestion and (ii) speedup due to structural properties of G2G^2. We demonstrate these two phenomena via two fundamental graph problems, namely, Minimum Vertex Cover (MVC) and Minimum Dominating Set (MDS). Among our many contributions, the highlights are the following. - In the CONGEST model, we show an O(n/ϵ)O(n/\epsilon)-round (1+ϵ)(1+\epsilon)-approximation algorithm for MVC on G2G^2, while no o(n2)o(n^2)-round algorithm is known for any better-than-2 approximation for MVC on GG. - We show a centralized polynomial time 5/35/3-approximation algorithm for MVC on G2G^2, whereas a better-than-2 approximation is UGC-hard for GG. - In contrast, for MDS, in the CONGEST model, we show an Ω~(n2)\tilde{\Omega}(n^2) lower bound for a constant approximation factor for MDS on G2G^2, whereas an Ω(n2)\Omega(n^2) lower bound for MDS on GG is known only for exact computation. In addition to these highlighted results, we prove a number of other results in the distributed CONGEST model including an Ω~(n2)\tilde{\Omega}(n^2) lower bound for computing an exact solution to MVC on G2G^2, a conditional hardness result for obtaining a (1+ϵ)(1+\epsilon)-approximation to MVC on G2G^2, and an O(logΔ)O(\log \Delta)-approximation to the MDS problem on G2G^2 in \mbox{poly}\log n rounds.Comment: Appears in PODC 2020. 40 pages, 7 figure

    Accelerated Gossip in Networks of Given Dimension using Jacobi Polynomial Iterations

    Get PDF
    Consider a network of agents connected by communication links, where each agent holds a real value. The gossip problem consists in estimating the average of the values diffused in the network in a distributed manner. We develop a method solving the gossip problem that depends only on the spectral dimension of the network, that is, in the communication network set-up, the dimension of the space in which the agents live. This contrasts with previous work that required the spectral gap of the network as a parameter, or suffered from slow mixing. Our method shows an important improvement over existing algorithms in the non-asymptotic regime, i.e., when the values are far from being fully mixed in the network. Our approach stems from a polynomial-based point of view on gossip algorithms, as well as an approximation of the spectral measure of the graphs with a Jacobi measure. We show the power of the approach with simulations on various graphs, and with performance guarantees on graphs of known spectral dimension, such as grids and random percolation bonds. An extension of this work to distributed Laplacian solvers is discussed. As a side result, we also use the polynomial-based point of view to show the convergence of the message passing algorithm for gossip of Moallemi \& Van Roy on regular graphs. The explicit computation of the rate of the convergence shows that message passing has a slow rate of convergence on graphs with small spectral gap

    Design of Self-Stabilizing Approximation Algorithms via a Primal-Dual Approach

    Get PDF
    Self-stabilization is an important concept in the realm of fault-tolerant distributed computing. In this paper, we propose a new approach that relies on the properties of linear programming duality to obtain self-stabilizing approximation algorithms for distributed graph optimization problems. The power of this new approach is demonstrated by the following results: - A self-stabilizing 2(1+?)-approximation algorithm for minimum weight vertex cover that converges in O(log? /(?log log ?)) synchronous rounds. - A self-stabilizing ?-approximation algorithm for maximum weight independent set that converges in O(?+log^* n) synchronous rounds. - A self-stabilizing ((2?+1)(1+?))-approximation algorithm for minimum weight dominating set in ?-arboricity graphs that converges in O((log?)/?) synchronous rounds. In all of the above, ? denotes the maximum degree. Our technique improves upon previous results in terms of time complexity while incurring only an additive O(log n) overhead to the message size. In addition, to the best of our knowledge, we provide the first self-stabilizing algorithms for the weighted versions of minimum vertex cover and maximum independent set

    Deterministic Distributed Algorithms and Lower Bounds in the Hybrid Model

    Get PDF
    The HYBRID model was recently introduced by Augustine et al. [John Augustine et al., 2020] in order to characterize from an algorithmic standpoint the capabilities of networks which combine multiple communication modes. Concretely, it is assumed that the standard LOCAL model of distributed computing is enhanced with the feature of all-to-all communication, but with very limited bandwidth, captured by the node-capacitated clique (NCC). In this work we provide several new insights on the power of hybrid networks for fundamental problems in distributed algorithms. First, we present a deterministic algorithm which solves any problem on a sparse n-node graph in ??(?n) rounds of HYBRID, where the notation ??(?) suppresses polylogarithmic factors of n. We combine this primitive with several sparsification techniques to obtain efficient distributed algorithms for general graphs. Most notably, for the all-pairs shortest paths problem we give deterministic (1 + ?)- and log n/log log n-approximate algorithms for unweighted and weighted graphs respectively with round complexity ??(?n) in HYBRID, closely matching the performance of the state of the art randomized algorithm of Kuhn and Schneider [Kuhn and Schneider, 2020]. Moreover, we make a connection with the Ghaffari-Haeupler framework of low-congestion shortcuts [Mohsen Ghaffari and Bernhard Haeupler, 2016], leading - among others - to a (1 + ?)-approximate algorithm for Min-Cut after ?(polylog (n)) rounds, with high probability, even if we restrict local edges to transfer ?(log n) bits per round. Finally, we prove via a reduction from the set disjointness problem that ??(n^{1/3}) rounds are required to determine the radius of an unweighted graph, as well as a (3/2 - ?)-approximation for weighted graphs. As a byproduct, we show an ??(n) round-complexity lower bound for computing a (4/3 - ?)-approximation of the radius in the broadcast variant of the congested clique, even for unweighted graphs

    Wireless Scheduling with Power Control

    Full text link
    We consider the scheduling of arbitrary wireless links in the physical model of interference to minimize the time for satisfying all requests. We study here the combined problem of scheduling and power control, where we seek both an assignment of power settings and a partition of the links so that each set satisfies the signal-to-interference-plus-noise (SINR) constraints. We give an algorithm that attains an approximation ratio of O(lognloglogΔ)O(\log n \cdot \log\log \Delta), where nn is the number of links and Δ\Delta is the ratio between the longest and the shortest link length. Under the natural assumption that lengths are represented in binary, this gives the first approximation ratio that is polylogarithmic in the size of the input. The algorithm has the desirable property of using an oblivious power assignment, where the power assigned to a sender depends only on the length of the link. We give evidence that this dependence on Δ\Delta is unavoidable, showing that any reasonably-behaving oblivious power assignment results in a Ω(loglogΔ)\Omega(\log\log \Delta)-approximation. These results hold also for the (weighted) capacity problem of finding a maximum (weighted) subset of links that can be scheduled in a single time slot. In addition, we obtain improved approximation for a bidirectional variant of the scheduling problem, give partial answers to questions about the utility of graphs for modeling physical interference, and generalize the setting from the standard 2-dimensional Euclidean plane to doubling metrics. Finally, we explore the utility of graph models in capturing wireless interference.Comment: Revised full versio

    Message and time efficient multi-broadcast schemes

    Full text link
    We consider message and time efficient broadcasting and multi-broadcasting in wireless ad-hoc networks, where a subset of nodes, each with a unique rumor, wish to broadcast their rumors to all destinations while minimizing the total number of transmissions and total time until all rumors arrive to their destination. Under centralized settings, we introduce a novel approximation algorithm that provides almost optimal results with respect to the number of transmissions and total time, separately. Later on, we show how to efficiently implement this algorithm under distributed settings, where the nodes have only local information about their surroundings. In addition, we show multiple approximation techniques based on the network collision detection capabilities and explain how to calibrate the algorithms' parameters to produce optimal results for time and messages.Comment: In Proceedings FOMC 2013, arXiv:1310.459

    Distributed Estimation and Control of Algebraic Connectivity over Random Graphs

    Full text link
    In this paper we propose a distributed algorithm for the estimation and control of the connectivity of ad-hoc networks in the presence of a random topology. First, given a generic random graph, we introduce a novel stochastic power iteration method that allows each node to estimate and track the algebraic connectivity of the underlying expected graph. Using results from stochastic approximation theory, we prove that the proposed method converges almost surely (a.s.) to the desired value of connectivity even in the presence of imperfect communication scenarios. The estimation strategy is then used as a basic tool to adapt the power transmitted by each node of a wireless network, in order to maximize the network connectivity in the presence of realistic Medium Access Control (MAC) protocols or simply to drive the connectivity toward a desired target value. Numerical results corroborate our theoretical findings, thus illustrating the main features of the algorithm and its robustness to fluctuations of the network graph due to the presence of random link failures.Comment: To appear in IEEE Transactions on Signal Processin
    corecore