12,121 research outputs found

    Structuring Unreliable Radio Networks

    Get PDF
    In this paper we study the problem of building a connected dominating set with constant degree (CCDS) in the dual graph radio network model. This model includes two types of links: reliable links, which always deliver messages, and unreliable links, which sometimes fail to deliver messages. Real networks compensate for this differing quality by deploying low-layer detection protocols to filter unreliable from reliable links. With this in mind, we begin by presenting an algorithm that solves the CCDS problem in the dual graph model under the assumption that every process u is provided with a local "link detector set" consisting of every neighbor connected to u by a reliable link. The algorithm solves the CCDS problem in O((Delta log2(n)/b) + log3(n)) rounds, with high probability, where Delta is the maximum degree in the reliable link graph, n is the network size, and b is an upper bound in bits on the message size. The algorithm works by first building a Maximal Independent Set (MIS) in log3(n) time, and then leveraging the local topology knowledge to efficiently connect nearby MIS processes. A natural follow up question is whether the link detector must be perfectly reliable to solve the CCDS problem. To answer this question, we first describe an algorithm that builds a CCDS in O(Delta polylog(n)) time under the assumption of O(1) unreliable links included in each link detector set. We then prove this algorithm to be (almost) tight by showing that the possible inclusion of only a single unreliable link in each process's local link detector set is sufficient to require Omega(Delta) rounds to solve the CCDS problem, regardless of message size. We conclude by discussing how to apply our algorithm in the setting where the topology of reliable and unreliable links can change over time

    Structuring Unreliable Radio Networks

    Get PDF
    In this paper we study the problem of building a connected dominating set with constant degree (CCDS) in the dual graph radio network model [4,9,10]. This model includes two types of links: reliable, which always deliver messages, and unreliable, which sometimes fail to deliver messages. Real networks compensate for this differing quality by deploying low-layer detection protocols to filter unreliable from reliable links. With this in mind, we begin by presenting an algorithm that solves the CCDS problem in the dual graph model under the assumption that every process u is provided a local link detector set consisting of every neighbor connected to u by a reliable link. The algorithm solves the CCDS problem in O(Δ\log[superscript 2] n/b + log[superscript 3] n) rounds, with high probability, where Δ is the maximum degree in the reliable link graph, n is the network size, and b is an upper bound in bits on the message size. The algorithm works by first building a Maximal Independent Set (MIS) in log[superscript 3] n time, and then leveraging the local topology knowledge to efficiently connect nearby MIS processes. A natural follow up question is whether the link detector must be perfectly reliable to solve the CCDS problem. With this in mind, we first describe an algorithm that builds a CCDS in O(Δpolylog(n)) time under the assumption of O(1) unreliable links included in each link detector set. We then prove this algorithm to be (almost) tight by showing that the possible inclusion of only a single unreliable link in each process's local link detector set is sufficient to require Ω(Δ) rounds to solve the CCDS problem, regardless of message size. We conclude by discussing how to apply our algorithm in the setting where the topology of reliable and unreliable links can change over time.Simons Foundation. (Postdoctoral Fellows Program)United States. Air Force Office of Scientific Research (Award FA9550-08-1-0159)National Science Foundation (U.S.) (Award CCF-0937274)National Science Foundation (U.S.) (Award CCF-0726514)National Science Foundation (U.S.) (Purdue University) (Science and Technology Center Award 0939370-CCF

    Lower Bounds for Structuring Unreliable Radio Networks

    Full text link
    In this paper, we study lower bounds for randomized solutions to the maximal independent set (MIS) and connected dominating set (CDS) problems in the dual graph model of radio networks---a generalization of the standard graph-based model that now includes unreliable links controlled by an adversary. We begin by proving that a natural geographic constraint on the network topology is required to solve these problems efficiently (i.e., in time polylogarthmic in the network size). We then prove the importance of the assumption that nodes are provided advance knowledge of their reliable neighbors (i.e, neighbors connected by reliable links). Combined, these results answer an open question by proving that the efficient MIS and CDS algorithms from [Censor-Hillel, PODC 2011] are optimal with respect to their dual graph model assumptions. They also provide insight into what properties of an unreliable network enable efficient local computation.Comment: An extended abstract of this work appears in the 2014 proceedings of the International Symposium on Distributed Computing (DISC

    Bounds on Contention Management in Radio Networks

    Get PDF
    The local broadcast problem assumes that processes in a wireless network are provided messages, one by one, that must be delivered to their neighbors. In this paper, we prove tight bounds for this problem in two well-studied wireless network models: the classical model, in which links are reliable and collisions consistent, and the more recent dual graph model, which introduces unreliable edges. Our results prove that the Decay strategy, commonly used for local broadcast in the classical setting, is optimal. They also establish a separation between the two models, proving that the dual graph setting is strictly harder than the classical setting, with respect to this primitive

    Broadcasting in Noisy Radio Networks

    Full text link
    The widely-studied radio network model [Chlamtac and Kutten, 1985] is a graph-based description that captures the inherent impact of collisions in wireless communication. In this model, the strong assumption is made that node vv receives a message from a neighbor if and only if exactly one of its neighbors broadcasts. We relax this assumption by introducing a new noisy radio network model in which random faults occur at senders or receivers. Specifically, for a constant noise parameter p[0,1)p \in [0,1), either every sender has probability pp of transmitting noise or every receiver of a single transmission in its neighborhood has probability pp of receiving noise. We first study single-message broadcast algorithms in noisy radio networks and show that the Decay algorithm [Bar-Yehuda et al., 1992] remains robust in the noisy model while the diameter-linear algorithm of Gasieniec et al., 2007 does not. We give a modified version of the algorithm of Gasieniec et al., 2007 that is robust to sender and receiver faults, and extend both this modified algorithm and the Decay algorithm to robust multi-message broadcast algorithms. We next investigate the extent to which (network) coding improves throughput in noisy radio networks. We address the previously perplexing result of Alon et al. 2014 that worst case coding throughput is no better than worst case routing throughput up to constants: we show that the worst case throughput performance of coding is, in fact, superior to that of routing -- by a Θ(log(n))\Theta(\log(n)) gap -- provided receiver faults are introduced. However, we show that any coding or routing scheme for the noiseless setting can be transformed to be robust to sender faults with only a constant throughput overhead. These transformations imply that the results of Alon et al., 2014 carry over to noisy radio networks with sender faults.Comment: Principles of Distributed Computing 201

    The Cost of Global Broadcast in Dynamic Radio Networks

    Get PDF
    We study the single-message broadcast problem in dynamic radio networks. We show that the time complexity of the problem depends on the amount of stability and connectivity of the dynamic network topology and on the adaptiveness of the adversary providing the dynamic topology. More formally, we model communication using the standard graph-based radio network model. To model the dynamic network, we use a generalization of the synchronous dynamic graph model introduced in [Kuhn et al., STOC 2010]. For integer parameters T1T\geq 1 and k1k\geq 1, we call a dynamic graph TT-interval kk-connected if for every interval of TT consecutive rounds, there exists a kk-vertex-connected stable subgraph. Further, for an integer parameter τ0\tau\geq 0, we say that the adversary providing the dynamic network is τ\tau-oblivious if for constructing the graph of some round tt, the adversary has access to all the randomness (and states) of the algorithm up to round tτt-\tau. As our main result, we show that for any T1T\geq 1, any k1k\geq 1, and any τ1\tau\geq 1, for a τ\tau-oblivious adversary, there is a distributed algorithm to broadcast a single message in time O((1+nkmin{τ,T})nlog3n)O\big(\big(1+\frac{n}{k\cdot\min\left\{\tau,T\right\}}\big)\cdot n\log^3 n\big). We further show that even for large interval kk-connectivity, efficient broadcast is not possible for the usual adaptive adversaries. For a 11-oblivious adversary, we show that even for any T(n/k)1εT\leq (n/k)^{1-\varepsilon} (for any constant ε>0\varepsilon>0) and for any k1k\geq 1, global broadcast in TT-interval kk-connected networks requires at least Ω(n2/(k2logn))\Omega(n^2/(k^2\log n)) time. Further, for a 00 oblivious adversary, broadcast cannot be solved in TT-interval kk-connected networks as long as T<nkT<n-k.Comment: 17 pages, conference version appeared in OPODIS 201

    Binding time: Harold Innis and the balance of new media

    Get PDF
    Much has been made of the impacts of digital media on the experience of space: new modes of perception and action at a distance: accelerating globalisation; shifting boundaries between work and home life; and so on. It is less common to read about the impacts of digital media on the experience of time. Yet, the digitisation of cultural practices and artefacts has significant implications for structuring our relationships with both the future and the past. In the theoretical traditions concerned with technology and time, the work of Harold Innis, a Canadian economist and communications theorist, offers an approach to understanding the social significance of all kinds of media. He analysed how different media relate to space and time: space-binding media extend influence and meanings over distances, helping to build empires and develop cohesion across space; while time-binding media influence cultural patterns in duration. For Innis, civilisations can be measured by their balance between managing time and controlling space. If this remains the case today, how has the computer changed this balance in our own culture? This paper examines the extent to which Innis’s concepts about media still apply today

    A (Truly) Local Broadcast Layer for Unreliable Radio Networks

    Get PDF
    In this paper, we implement an efficient local broadcast service for the dual graph model, which describes communication in a radio network with both reliable and unreliable links. Our local broadcast service offers probabilistic latency guarantees for: (1) message delivery to all reliable neighbors (i.e., neighbors connected by reliable links), and (2) receiving some message when one or more reliable neighbors are broadcasting. This service significantly simplifies the design and analysis of algorithms for the otherwise challenging dual graph model. To this end, we also note that our solution can be interpreted as an implementation of the abstract MAC layer specification---therefore translating the growing corpus of algorithmic results studied on top of this layer to the dual graph model. At the core of our service is a seed agreement routine which enables nodes in the network to achieve "good enough" coordination to overcome the difficulties of unpredictable link behavior. Because this routine has potential application to other problems in this setting, we capture it with a formal specification---simplifying its reuse in other algorithms. Finally, we note that in a break from much work on distributed radio network algorithms, our problem definitions (including error bounds), implementation, and analysis do not depend on global network parameters such as the network size, a goal which required new analysis techniques. We argue that breaking the dependence of these algorithms on global parameters makes more sense and aligns better with the rise of ubiquitous computing, where devices will be increasingly working locally in an otherwise massive network. Our push for locality, in other words, is a contribution independent of the specific radio network model and problem studied here
    corecore