16,764 research outputs found
Secure Partial Repair in Wireless Caching Networks with Broadcast Channels
We study security in partial repair in wireless caching networks where parts
of the stored packets in the caching nodes are susceptible to be erased. Let us
denote a caching node that has lost parts of its stored packets as a sick
caching node and a caching node that has not lost any packet as a healthy
caching node. In partial repair, a set of caching nodes (among sick and healthy
caching nodes) broadcast information to other sick caching nodes to recover the
erased packets. The broadcast information from a caching node is assumed to be
received without any error by all other caching nodes. All the sick caching
nodes then are able to recover their erased packets, while using the broadcast
information and the nonerased packets in their storage as side information. In
this setting, if an eavesdropper overhears the broadcast channels, it might
obtain some information about the stored file. We thus study secure partial
repair in the senses of information-theoretically strong and weak security. In
both senses, we investigate the secrecy caching capacity, namely, the maximum
amount of information which can be stored in the caching network such that
there is no leakage of information during a partial repair process. We then
deduce the strong and weak secrecy caching capacities, and also derive the
sufficient finite field sizes for achieving the capacities. Finally, we propose
optimal secure codes for exact partial repair, in which the recovered packets
are exactly the same as erased packets.Comment: To Appear in IEEE Conference on Communication and Network Security
(CNS
Fundamentals of Large Sensor Networks: Connectivity, Capacity, Clocks and Computation
Sensor networks potentially feature large numbers of nodes that can sense
their environment over time, communicate with each other over a wireless
network, and process information. They differ from data networks in that the
network as a whole may be designed for a specific application. We study the
theoretical foundations of such large scale sensor networks, addressing four
fundamental issues- connectivity, capacity, clocks and function computation.
To begin with, a sensor network must be connected so that information can
indeed be exchanged between nodes. The connectivity graph of an ad-hoc network
is modeled as a random graph and the critical range for asymptotic connectivity
is determined, as well as the critical number of neighbors that a node needs to
connect to. Next, given connectivity, we address the issue of how much data can
be transported over the sensor network. We present fundamental bounds on
capacity under several models, as well as architectural implications for how
wireless communication should be organized.
Temporal information is important both for the applications of sensor
networks as well as their operation.We present fundamental bounds on the
synchronizability of clocks in networks, and also present and analyze
algorithms for clock synchronization. Finally we turn to the issue of gathering
relevant information, that sensor networks are designed to do. One needs to
study optimal strategies for in-network aggregation of data, in order to
reliably compute a composite function of sensor measurements, as well as the
complexity of doing so. We address the issue of how such computation can be
performed efficiently in a sensor network and the algorithms for doing so, for
some classes of functions.Comment: 10 pages, 3 figures, Submitted to the Proceedings of the IEE
Communication Over a Wireless Network With Random Connections
A network of nodes in which pairs communicate over a shared wireless medium is analyzed. We consider the maximum total aggregate traffic flow possible as given by the number of users multiplied by their data rate. The model in this paper differs substantially from the many existing approaches in that the channel connections in this network are entirely random: rather than being governed by geometry and a decay-versus-distance law, the strengths of the connections between nodes are drawn independently from a common distribution. Such a model is appropriate for environments where the first-order effect that governs the signal strength at a receiving node is a random event (such as the existence of an obstacle), rather than the distance from the transmitter. It is shown that the aggregate traffic flow as a function of the number of nodes n is a strong function of the channel distribution. In particular, for certain distributions the aggregate traffic flow is at least n/(log n)^d for some d≫0, which is significantly larger than the O(sqrt n) results obtained for many geometric models. The results provide guidelines for the connectivity that is needed for large aggregate traffic. The relation between the proposed model and existing distance-based models is shown in some cases
Towards a System Theoretic Approach to Wireless Network Capacity in Finite Time and Space
In asymptotic regimes, both in time and space (network size), the derivation
of network capacity results is grossly simplified by brushing aside queueing
behavior in non-Jackson networks. This simplifying double-limit model, however,
lends itself to conservative numerical results in finite regimes. To properly
account for queueing behavior beyond a simple calculus based on average rates,
we advocate a system theoretic methodology for the capacity problem in finite
time and space regimes. This methodology also accounts for spatial correlations
arising in networks with CSMA/CA scheduling and it delivers rigorous
closed-form capacity results in terms of probability distributions. Unlike
numerous existing asymptotic results, subject to anecdotal practical concerns,
our transient one can be used in practical settings: for example, to compute
the time scales at which multi-hop routing is more advantageous than single-hop
routing
On the Catalyzing Effect of Randomness on the Per-Flow Throughput in Wireless Networks
This paper investigates the throughput capacity of a flow crossing a
multi-hop wireless network, whose geometry is characterized by general
randomness laws including Uniform, Poisson, Heavy-Tailed distributions for both
the nodes' densities and the number of hops. The key contribution is to
demonstrate \textit{how} the \textit{per-flow throughput} depends on the
distribution of 1) the number of nodes inside hops' interference sets, 2)
the number of hops , and 3) the degree of spatial correlations. The
randomness in both 's and is advantageous, i.e., it can yield larger
scalings (as large as ) than in non-random settings. An interesting
consequence is that the per-flow capacity can exhibit the opposite behavior to
the network capacity, which was shown to suffer from a logarithmic decrease in
the presence of randomness. In turn, spatial correlations along the end-to-end
path are detrimental by a logarithmic term
Towards a Simple Relationship to Estimate the Capacity of Static and Mobile Wireless Networks
Extensive research has been done on studying the capacity of wireless
multi-hop networks. These efforts have led to many sophisticated and customized
analytical studies on the capacity of particular networks. While most of the
analyses are intellectually challenging, they lack universal properties that
can be extended to study the capacity of a different network. In this paper, we
sift through various capacity-impacting parameters and present a simple
relationship that can be used to estimate the capacity of both static and
mobile networks. Specifically, we show that the network capacity is determined
by the average number of simultaneous transmissions, the link capacity and the
average number of transmissions required to deliver a packet to its
destination. Our result is valid for both finite networks and asymptotically
infinite networks. We then use this result to explain and better understand the
insights of some existing results on the capacity of static networks, mobile
networks and hybrid networks and the multicast capacity. The capacity analysis
using the aforementioned relationship often becomes simpler. The relationship
can be used as a powerful tool to estimate the capacity of different networks.
Our work makes important contributions towards developing a generic methodology
for network capacity analysis that is applicable to a variety of different
scenarios.Comment: accepted to appear in IEEE Transactions on Wireless Communication
Capacity of wireless erasure networks
In this paper, a special class of wireless networks, called wireless erasure networks, is considered. In these networks, each node is connected to a set of nodes by possibly correlated erasure channels. The network model incorporates the broadcast nature of the wireless environment by requiring each node to send the same signal on all outgoing channels. However, we assume there is no interference in reception. Such models are therefore appropriate for wireless networks where all information transmission is packetized and where some mechanism for interference avoidance is already built in. This paper looks at multicast problems over these networks. The capacity under the assumption that erasure locations on all the links of the network are provided to the destinations is obtained. It turns out that the capacity region has a nice max-flow min-cut interpretation. The definition of cut-capacity in these networks incorporates the broadcast property of the wireless medium. It is further shown that linear coding at nodes in the network suffices to achieve the capacity region. Finally, the performance of different coding schemes in these networks when no side information is available to the destinations is analyzed
Beyond Geometry : Towards Fully Realistic Wireless Models
Signal-strength models of wireless communications capture the gradual fading
of signals and the additivity of interference. As such, they are closer to
reality than other models. However, nearly all theoretic work in the SINR model
depends on the assumption of smooth geometric decay, one that is true in free
space but is far off in actual environments. The challenge is to model
realistic environments, including walls, obstacles, reflections and anisotropic
antennas, without making the models algorithmically impractical or analytically
intractable.
We present a simple solution that allows the modeling of arbitrary static
situations by moving from geometry to arbitrary decay spaces. The complexity of
a setting is captured by a metricity parameter Z that indicates how far the
decay space is from satisfying the triangular inequality. All results that hold
in the SINR model in general metrics carry over to decay spaces, with the
resulting time complexity and approximation depending on Z in the same way that
the original results depends on the path loss term alpha. For distributed
algorithms, that to date have appeared to necessarily depend on the planarity,
we indicate how they can be adapted to arbitrary decay spaces.
Finally, we explore the dependence on Z in the approximability of core
problems. In particular, we observe that the capacity maximization problem has
exponential upper and lower bounds in terms of Z in general decay spaces. In
Euclidean metrics and related growth-bounded decay spaces, the performance
depends on the exact metricity definition, with a polynomial upper bound in
terms of Z, but an exponential lower bound in terms of a variant parameter phi.
On the plane, the upper bound result actually yields the first approximation of
a capacity-type SINR problem that is subexponential in alpha
- …