16,764 research outputs found

    Secure Partial Repair in Wireless Caching Networks with Broadcast Channels

    Full text link
    We study security in partial repair in wireless caching networks where parts of the stored packets in the caching nodes are susceptible to be erased. Let us denote a caching node that has lost parts of its stored packets as a sick caching node and a caching node that has not lost any packet as a healthy caching node. In partial repair, a set of caching nodes (among sick and healthy caching nodes) broadcast information to other sick caching nodes to recover the erased packets. The broadcast information from a caching node is assumed to be received without any error by all other caching nodes. All the sick caching nodes then are able to recover their erased packets, while using the broadcast information and the nonerased packets in their storage as side information. In this setting, if an eavesdropper overhears the broadcast channels, it might obtain some information about the stored file. We thus study secure partial repair in the senses of information-theoretically strong and weak security. In both senses, we investigate the secrecy caching capacity, namely, the maximum amount of information which can be stored in the caching network such that there is no leakage of information during a partial repair process. We then deduce the strong and weak secrecy caching capacities, and also derive the sufficient finite field sizes for achieving the capacities. Finally, we propose optimal secure codes for exact partial repair, in which the recovered packets are exactly the same as erased packets.Comment: To Appear in IEEE Conference on Communication and Network Security (CNS

    Fundamentals of Large Sensor Networks: Connectivity, Capacity, Clocks and Computation

    Full text link
    Sensor networks potentially feature large numbers of nodes that can sense their environment over time, communicate with each other over a wireless network, and process information. They differ from data networks in that the network as a whole may be designed for a specific application. We study the theoretical foundations of such large scale sensor networks, addressing four fundamental issues- connectivity, capacity, clocks and function computation. To begin with, a sensor network must be connected so that information can indeed be exchanged between nodes. The connectivity graph of an ad-hoc network is modeled as a random graph and the critical range for asymptotic connectivity is determined, as well as the critical number of neighbors that a node needs to connect to. Next, given connectivity, we address the issue of how much data can be transported over the sensor network. We present fundamental bounds on capacity under several models, as well as architectural implications for how wireless communication should be organized. Temporal information is important both for the applications of sensor networks as well as their operation.We present fundamental bounds on the synchronizability of clocks in networks, and also present and analyze algorithms for clock synchronization. Finally we turn to the issue of gathering relevant information, that sensor networks are designed to do. One needs to study optimal strategies for in-network aggregation of data, in order to reliably compute a composite function of sensor measurements, as well as the complexity of doing so. We address the issue of how such computation can be performed efficiently in a sensor network and the algorithms for doing so, for some classes of functions.Comment: 10 pages, 3 figures, Submitted to the Proceedings of the IEE

    Communication Over a Wireless Network With Random Connections

    Get PDF
    A network of nodes in which pairs communicate over a shared wireless medium is analyzed. We consider the maximum total aggregate traffic flow possible as given by the number of users multiplied by their data rate. The model in this paper differs substantially from the many existing approaches in that the channel connections in this network are entirely random: rather than being governed by geometry and a decay-versus-distance law, the strengths of the connections between nodes are drawn independently from a common distribution. Such a model is appropriate for environments where the first-order effect that governs the signal strength at a receiving node is a random event (such as the existence of an obstacle), rather than the distance from the transmitter. It is shown that the aggregate traffic flow as a function of the number of nodes n is a strong function of the channel distribution. In particular, for certain distributions the aggregate traffic flow is at least n/(log n)^d for some d≫0, which is significantly larger than the O(sqrt n) results obtained for many geometric models. The results provide guidelines for the connectivity that is needed for large aggregate traffic. The relation between the proposed model and existing distance-based models is shown in some cases

    Towards a System Theoretic Approach to Wireless Network Capacity in Finite Time and Space

    Get PDF
    In asymptotic regimes, both in time and space (network size), the derivation of network capacity results is grossly simplified by brushing aside queueing behavior in non-Jackson networks. This simplifying double-limit model, however, lends itself to conservative numerical results in finite regimes. To properly account for queueing behavior beyond a simple calculus based on average rates, we advocate a system theoretic methodology for the capacity problem in finite time and space regimes. This methodology also accounts for spatial correlations arising in networks with CSMA/CA scheduling and it delivers rigorous closed-form capacity results in terms of probability distributions. Unlike numerous existing asymptotic results, subject to anecdotal practical concerns, our transient one can be used in practical settings: for example, to compute the time scales at which multi-hop routing is more advantageous than single-hop routing

    On the Catalyzing Effect of Randomness on the Per-Flow Throughput in Wireless Networks

    Get PDF
    This paper investigates the throughput capacity of a flow crossing a multi-hop wireless network, whose geometry is characterized by general randomness laws including Uniform, Poisson, Heavy-Tailed distributions for both the nodes' densities and the number of hops. The key contribution is to demonstrate \textit{how} the \textit{per-flow throughput} depends on the distribution of 1) the number of nodes NjN_j inside hops' interference sets, 2) the number of hops KK, and 3) the degree of spatial correlations. The randomness in both NjN_j's and KK is advantageous, i.e., it can yield larger scalings (as large as Θ(n)\Theta(n)) than in non-random settings. An interesting consequence is that the per-flow capacity can exhibit the opposite behavior to the network capacity, which was shown to suffer from a logarithmic decrease in the presence of randomness. In turn, spatial correlations along the end-to-end path are detrimental by a logarithmic term

    Towards a Simple Relationship to Estimate the Capacity of Static and Mobile Wireless Networks

    Full text link
    Extensive research has been done on studying the capacity of wireless multi-hop networks. These efforts have led to many sophisticated and customized analytical studies on the capacity of particular networks. While most of the analyses are intellectually challenging, they lack universal properties that can be extended to study the capacity of a different network. In this paper, we sift through various capacity-impacting parameters and present a simple relationship that can be used to estimate the capacity of both static and mobile networks. Specifically, we show that the network capacity is determined by the average number of simultaneous transmissions, the link capacity and the average number of transmissions required to deliver a packet to its destination. Our result is valid for both finite networks and asymptotically infinite networks. We then use this result to explain and better understand the insights of some existing results on the capacity of static networks, mobile networks and hybrid networks and the multicast capacity. The capacity analysis using the aforementioned relationship often becomes simpler. The relationship can be used as a powerful tool to estimate the capacity of different networks. Our work makes important contributions towards developing a generic methodology for network capacity analysis that is applicable to a variety of different scenarios.Comment: accepted to appear in IEEE Transactions on Wireless Communication

    Capacity of wireless erasure networks

    Get PDF
    In this paper, a special class of wireless networks, called wireless erasure networks, is considered. In these networks, each node is connected to a set of nodes by possibly correlated erasure channels. The network model incorporates the broadcast nature of the wireless environment by requiring each node to send the same signal on all outgoing channels. However, we assume there is no interference in reception. Such models are therefore appropriate for wireless networks where all information transmission is packetized and where some mechanism for interference avoidance is already built in. This paper looks at multicast problems over these networks. The capacity under the assumption that erasure locations on all the links of the network are provided to the destinations is obtained. It turns out that the capacity region has a nice max-flow min-cut interpretation. The definition of cut-capacity in these networks incorporates the broadcast property of the wireless medium. It is further shown that linear coding at nodes in the network suffices to achieve the capacity region. Finally, the performance of different coding schemes in these networks when no side information is available to the destinations is analyzed

    Beyond Geometry : Towards Fully Realistic Wireless Models

    Full text link
    Signal-strength models of wireless communications capture the gradual fading of signals and the additivity of interference. As such, they are closer to reality than other models. However, nearly all theoretic work in the SINR model depends on the assumption of smooth geometric decay, one that is true in free space but is far off in actual environments. The challenge is to model realistic environments, including walls, obstacles, reflections and anisotropic antennas, without making the models algorithmically impractical or analytically intractable. We present a simple solution that allows the modeling of arbitrary static situations by moving from geometry to arbitrary decay spaces. The complexity of a setting is captured by a metricity parameter Z that indicates how far the decay space is from satisfying the triangular inequality. All results that hold in the SINR model in general metrics carry over to decay spaces, with the resulting time complexity and approximation depending on Z in the same way that the original results depends on the path loss term alpha. For distributed algorithms, that to date have appeared to necessarily depend on the planarity, we indicate how they can be adapted to arbitrary decay spaces. Finally, we explore the dependence on Z in the approximability of core problems. In particular, we observe that the capacity maximization problem has exponential upper and lower bounds in terms of Z in general decay spaces. In Euclidean metrics and related growth-bounded decay spaces, the performance depends on the exact metricity definition, with a polynomial upper bound in terms of Z, but an exponential lower bound in terms of a variant parameter phi. On the plane, the upper bound result actually yields the first approximation of a capacity-type SINR problem that is subexponential in alpha
    corecore