1,519 research outputs found

    Detection Performance in Balanced Binary Relay Trees with Node and Link Failures

    Full text link
    We study the distributed detection problem in the context of a balanced binary relay tree, where the leaves of the tree correspond to NN identical and independent sensors generating binary messages. The root of the tree is a fusion center making an overall decision. Every other node is a relay node that aggregates the messages received from its child nodes into a new message and sends it up toward the fusion center. We derive upper and lower bounds for the total error probability PNP_N as explicit functions of NN in the case where nodes and links fail with certain probabilities. These characterize the asymptotic decay rate of the total error probability as NN goes to infinity. Naturally, this decay rate is not larger than that in the non-failure case, which is N\sqrt N. However, we derive an explicit necessary and sufficient condition on the decay rate of the local failure probabilities pkp_k (combination of node and link failure probabilities at each level) such that the decay rate of the total error probability in the failure case is the same as that of the non-failure case. More precisely, we show that logPN1=Θ(N)\log P_N^{-1}=\Theta(\sqrt N) if and only if logpk1=Ω(2k/2)\log p_k^{-1}=\Omega(2^{k/2})

    Submodularity and Optimality of Fusion Rules in Balanced Binary Relay Trees

    Full text link
    We study the distributed detection problem in a balanced binary relay tree, where the leaves of the tree are sensors generating binary messages. The root of the tree is a fusion center that makes the overall decision. Every other node in the tree is a fusion node that fuses two binary messages from its child nodes into a new binary message and sends it to the parent node at the next level. We assume that the fusion nodes at the same level use the same fusion rule. We call a string of fusion rules used at different levels a fusion strategy. We consider the problem of finding a fusion strategy that maximizes the reduction in the total error probability between the sensors and the fusion center. We formulate this problem as a deterministic dynamic program and express the solution in terms of Bellman's equations. We introduce the notion of stringsubmodularity and show that the reduction in the total error probability is a stringsubmodular function. Consequentially, we show that the greedy strategy, which only maximizes the level-wise reduction in the total error probability, is within a factor of the optimal strategy in terms of reduction in the total error probability

    Distributed Detection and Estimation in Wireless Sensor Networks

    Full text link
    In this article we consider the problems of distributed detection and estimation in wireless sensor networks. In the first part, we provide a general framework aimed to show how an efficient design of a sensor network requires a joint organization of in-network processing and communication. Then, we recall the basic features of consensus algorithm, which is a basic tool to reach globally optimal decisions through a distributed approach. The main part of the paper starts addressing the distributed estimation problem. We show first an entirely decentralized approach, where observations and estimations are performed without the intervention of a fusion center. Then, we consider the case where the estimation is performed at a fusion center, showing how to allocate quantization bits and transmit powers in the links between the nodes and the fusion center, in order to accommodate the requirement on the maximum estimation variance, under a constraint on the global transmit power. We extend the approach to the detection problem. Also in this case, we consider the distributed approach, where every node can achieve a globally optimal decision, and the case where the decision is taken at a central node. In the latter case, we show how to allocate coding bits and transmit power in order to maximize the detection probability, under constraints on the false alarm rate and the global transmit power. Then, we generalize consensus algorithms illustrating a distributed procedure that converges to the projection of the observation vector onto a signal subspace. We then address the issue of energy consumption in sensor networks, thus showing how to optimize the network topology in order to minimize the energy necessary to achieve a global consensus. Finally, we address the problem of matching the topology of the network to the graph describing the statistical dependencies among the observed variables.Comment: 92 pages, 24 figures. To appear in E-Reference Signal Processing, R. Chellapa and S. Theodoridis, Eds., Elsevier, 201

    Network coding for reliable wireless sensor networks

    Get PDF
    Wireless sensor networks are used in many applications and are now a key element in the increasingly growing Internet of Things. These networks are composed of small nodes including wireless communication modules, and in most of the cases are able to autonomously con gure themselves into networks, to ensure sensed data delivery. As more and more sensor nodes and networks join the Internet of Things, collaboration between geographically distributed systems are expected. Peer to peer overlay networks can assist in the federation of these systems, for them to collaborate. Since participating peers/proxies contribute to storage and processing, there is no burden on speci c servers and bandwidth bottlenecks are avoided. Network coding can be used to improve the performance of wireless sensor networks. The idea is for data from multiple links to be combined at intermediate encoding nodes, before further transmission. This technique proved to have a lot of potential in a wide range of applications. In the particular case of sensor networks, network coding based protocols and algorithms try to achieve a balance between low packet error rate and energy consumption. For network coding based constrained networks to be federated using peer to peer overlays, it is necessary to enable the storage of encoding vectors and coded data by such distributed storage systems. Packets can arrive to the overlay through any gateway/proxy (peers in the overlay), and lost packets can be recovered by the overlay (or client) using original and coded data that has been stored. The decoding process requires a decoding service at the overlay network. Such architecture, which is the focus of this thesis, will allow constrained networks to reduce packet error rate in an energy e cient way, while bene ting from an e ective distributed storage solution for their federation. This will serve as a basis for the proposal of mathematical models and algorithms that determine the most e ective routing trees, for packet forwarding toward sink/gateway nodes, and best amount and placement of encoding nodes.As redes de sensores sem fios são usadas em muitas aplicações e são hoje consideradas um elemento-chave para o desenvolvimento da Internet das Coisas. Compostas por nós de pequena dimensão que incorporam módulos de comunicação sem fios, grande parte destas redes possuem a capacidade de se configurarem de forma autónoma, formando sistemas em rede para garantir a entrega dos dados recolhidos. (…

    Performance optimization of wireless sensor networks for remote monitoring

    Get PDF
    Wireless sensor networks (WSNs) have gained worldwide attention in recent years because of their great potential for a variety of applications such as hazardous environment exploration, military surveillance, habitat monitoring, seismic sensing, and so on. In this thesis we study the use of WSNs for remote monitoring, where a wireless sensor network is deployed in a remote region for sensing phenomena of interest while its data monitoring center is located in a metropolitan area that is geographically distant from the monitored region. This application scenario poses great challenges since such kind of monitoring is typically large scale and expected to be operational for a prolonged period without human involvement. Also, the long distance between the monitored region and the data monitoring center requires that the sensed data must be transferred by the employment of a third-party communication service, which incurs service costs. Existing methodologies for performance optimization of WSNs base on that both the sensor network and its data monitoring center are co-located, and therefore are no longer applicable to the remote monitoring scenario. Thus, developing new techniques and approaches for severely resource-constrained WSNs is desperately needed to maintain sustainable, unattended remote monitoring with low cost. Specifically, this thesis addresses the key issues and tackles problems in the deployment of WSNs for remote monitoring from the following aspects. To maximize the lifetime of large-scale monitoring, we deal with the energy consumption imbalance issue by exploring multiple sinks. We develop scalable algorithms which determine the optimal number of sinks needed and their locations, thereby dynamically identifying the energy bottlenecks and balancing the data relay workload throughout the network. We conduct experiments and the experimental results demonstrate that the proposed algorithms significantly prolong the network lifetime. To eliminate imbalance of energy consumption among sensor nodes, a complementary strategy is to introduce a mobile sink for data gathering. However, the limited communication time between the mobile sink and nodes results in that only part of sensed data will be collected and the rest will be lost, for which we propose the concept of monitoring quality with the exploration of sensed data correlation among nodes. We devise a heuristic for monitoring quality maximization, which schedules the sink to collect data from selected nodes, and uses the collected data to recover the missing ones. We study the performance of the proposed heuristic and validate its effectiveness in improving the monitoring quality. To strive for the fine trade-off between two performance metrics: throughput and cost, we investigate novel problems of minimizing cost with guaranteed throughput, and maximizing throughput with minimal cost. We develop approximation algorithms which find reliable data routing in the WSN and strategically balance workload on the sinks. We prove that the delivered solutions are fractional of the optimum. We finally conclude our work and discuss potential research topics which derive from the studies of this thesis

    Energy-efficient routing protocols in heterogeneous wireless sensor networks

    Get PDF
    Sensor networks feature low-cost sensor devices with wireless network capability, limited transmit power, resource constraints and limited battery energy. The usage of cheap and tiny wireless sensors will allow very large networks to be deployed at a feasible cost to provide a bridge between information systems and the physical world. Such large-scale deployments will require routing protocols that scale to large network sizes in an energy-efficient way. This thesis addresses the design of such network routing methods. A classification of existing routing protocols and the key factors in their design (i.e., hardware, topology, applications) provides the motivation for the new three-tier architecture for heterogeneous networks built upon a generic software framework (GSF). A range of new routing algorithms have hence been developed with the design goals of scalability and energy-efficient performance of network protocols. They are respectively TinyReg - a routing algorithm based on regular-graph theory, TSEP - topological stable election protocol, and GAAC - an evolutionary algorithm based on genetic algorithms and ant colony algorithms. The design principle of our routing algorithms is that shortening the distance between the cluster-heads and the sink in the network, will minimise energy consumption in order to extend the network lifetime, will achieve energy efficiency. Their performance has been evaluated by simulation in an extensive range of scenarios, and compared to existing algorithms. It is shown that the newly proposed algorithms allow long-term continuous data collection in large networks, offering greater network longevity than existing solutions. These results confirm the validity of the GSF as an architectural approach to the deployment of large wireless sensor networks

    Optimizing on-demand resource deployment for peer-assisted content delivery

    Full text link
    Increasingly, content delivery solutions leverage client resources in exchange for services in a pee-to-peer (P2P) fashion. Such peer-assisted service paradigm promises significant infrastructure cost reduction, but suffers from the unpredictability associated with client resources, which is often exhibited as an imbalance between the contribution and consumption of resources by clients. This imbalance hinders the ability to guarantee a minimum service fidelity of these services to clients especially for real-time applications where content can not be cached. In this thesis, we propose a novel architectural service model that enables the establishment of higher fidelity services through (1) coordinating the content delivery to efficiently utilize the available resources, and (2) leasing the least additional cloud resources, available through special nodes (angels) that join the service on-demand, and only if needed, to complement the scarce resources available through clients. While the proposed service model can be deployed in many settings, this thesis focuses on peer-assisted content delivery applications, in which the scarce resource is typically the upstream capacity of clients. We target three applications that require the delivery of real-time as opposed to stale content. The first application is bulk-synchronous transfer, in which the goal of the system is to minimize the maximum distribution time - the time it takes to deliver the content to all clients in a group. The second application is live video streaming, in which the goal of the system is to maintain a given streaming quality. The third application is Tor, the anonymous onion routing network, in which the goal of the system is to boost performance (increase throughput and reduce latency) throughout the network, and especially for clients running bandwidth-intensive applications. For each of the above applications, we develop analytical models that efficiently allocate the already available resources. They also efficiently allocate additional on-demand resource to achieve a certain level of service. Our analytical models and efficient constructions depend on some simplifying, yet impractical, assumptions. Thus, inspired by our models and constructions, we develop practical techniques that we incorporate into prototypical peer-assisted angel-enabled cloud services. We evaluate these techniques through simulation and/or implementation

    Optimizing on-demand resource deployment for peer-assisted content delivery (PhD thesis)

    Full text link
    Increasingly, content delivery solutions leverage client resources in exchange for service in a peer-to-peer (P2P) fashion. Such peer-assisted service paradigms promise significant infrastructure cost reduction, but suffer from the unpredictability associated with client resources, which is often exhibited as an imbalance between the contribution and consumption of resources by clients. This imbalance hinders the ability to guarantee a minimum service fidelity of these services to the clients. In this thesis, we propose a novel architectural service model that enables the establishment of higher fidelity services through (1) coordinating the content delivery to optimally utilize the available resources, and (2) leasing the least additional cloud resources, available through special nodes (angels) that join the service on-demand, and only if needed, to complement the scarce resources available through clients. While the proposed service model can be deployed in many settings, this thesis focuses on peer-assisted content delivery applications, in which the scarce resource is typically the uplink capacity of clients. We target three applications that require the delivery of fresh as opposed to stale content. The first application is bulk-synchronous transfer, in which the goal of the system is to minimize the maximum distribution time -- the time it takes to deliver the content to all clients in a group. The second application is live streaming, in which the goal of the system is to maintain a given streaming quality. The third application is Tor, the anonymous onion routing network, in which the goal of the system is to boost performance (increase throughput and reduce latency) throughout the network, and especially for bandwidth-intensive applications. For each of the above applications, we develop mathematical models that optimally allocate the already available resources. They also optimally allocate additional on-demand resource to achieve a certain level of service. Our analytical models and efficient constructions depend on some simplifying, yet impractical, assumptions. Thus, inspired by our models and constructions, we develop practical techniques that we incorporate into prototypical peer-assisted angel-enabled cloud services. We evaluate those techniques through simulation and/or implementation. (Major Advisor: Azer Bestavros
    corecore