9,810 research outputs found
Rate-Distortion Classification for Self-Tuning IoT Networks
Many future wireless sensor networks and the Internet of Things are expected
to follow a software defined paradigm, where protocol parameters and behaviors
will be dynamically tuned as a function of the signal statistics. New protocols
will be then injected as a software as certain events occur. For instance, new
data compressors could be (re)programmed on-the-fly as the monitored signal
type or its statistical properties change. We consider a lossy compression
scenario, where the application tolerates some distortion of the gathered
signal in return for improved energy efficiency. To reap the full benefits of
this paradigm, we discuss an automatic sensor profiling approach where the
signal class, and in particular the corresponding rate-distortion curve, is
automatically assessed using machine learning tools (namely, support vector
machines and neural networks). We show that this curve can be reliably
estimated on-the-fly through the computation of a small number (from ten to
twenty) of statistical features on time windows of a few hundreds samples
EZ-AG: Structure-free data aggregation in MANETs using push-assisted self-repelling random walks
This paper describes EZ-AG, a structure-free protocol for duplicate
insensitive data aggregation in MANETs. The key idea in EZ-AG is to introduce a
token that performs a self-repelling random walk in the network and aggregates
information from nodes when they are visited for the first time. A
self-repelling random walk of a token on a graph is one in which at each step,
the token moves to a neighbor that has been visited least often. While
self-repelling random walks visit all nodes in the network much faster than
plain random walks, they tend to slow down when most of the nodes are already
visited. In this paper, we show that a single step push phase at each node can
significantly speed up the aggregation and eliminate this slow down. By doing
so, EZ-AG achieves aggregation in only O(N) time and messages. In terms of
overhead, EZ-AG outperforms existing structure-free data aggregation by a
factor of at least log(N) and achieves the lower bound for aggregation message
overhead. We demonstrate the scalability and robustness of EZ-AG using ns-3
simulations in networks ranging from 100 to 4000 nodes under different mobility
models and node speeds. We also describe a hierarchical extension for EZ-AG
that can produce multi-resolution aggregates at each node using only O(NlogN)
messages, which is a poly-logarithmic factor improvement over existing
techniques
Machine Learning in Wireless Sensor Networks: Algorithms, Strategies, and Applications
Wireless sensor networks monitor dynamic environments that change rapidly
over time. This dynamic behavior is either caused by external factors or
initiated by the system designers themselves. To adapt to such conditions,
sensor networks often adopt machine learning techniques to eliminate the need
for unnecessary redesign. Machine learning also inspires many practical
solutions that maximize resource utilization and prolong the lifespan of the
network. In this paper, we present an extensive literature review over the
period 2002-2013 of machine learning methods that were used to address common
issues in wireless sensor networks (WSNs). The advantages and disadvantages of
each proposed algorithm are evaluated against the corresponding problem. We
also provide a comparative guide to aid WSN designers in developing suitable
machine learning solutions for their specific application challenges.Comment: Accepted for publication in IEEE Communications Surveys and Tutorial
Dependability in Aggregation by Averaging
Aggregation is an important building block of modern distributed
applications, allowing the determination of meaningful properties (e.g. network
size, total storage capacity, average load, majorities, etc.) that are used to
direct the execution of the system. However, the majority of the existing
aggregation algorithms exhibit relevant dependability issues, when prospecting
their use in real application environments. In this paper, we reveal some
dependability issues of aggregation algorithms based on iterative averaging
techniques, giving some directions to solve them. This class of algorithms is
considered robust (when compared to common tree-based approaches), being
independent from the used routing topology and providing an aggregation result
at all nodes. However, their robustness is strongly challenged and their
correctness often compromised, when changing the assumptions of their working
environment to more realistic ones. The correctness of this class of algorithms
relies on the maintenance of a fundamental invariant, commonly designated as
"mass conservation". We will argue that this main invariant is often broken in
practical settings, and that additional mechanisms and modifications are
required to maintain it, incurring in some degradation of the algorithms
performance. In particular, we discuss the behavior of three representative
algorithms Push-Sum Protocol, Push-Pull Gossip protocol and Distributed Random
Grouping under asynchronous and faulty (with message loss and node crashes)
environments. More specifically, we propose and evaluate two new versions of
the Push-Pull Gossip protocol, which solve its message interleaving problem
(evidenced even in a synchronous operation mode).Comment: 14 pages. Presented in Inforum 200
The age of information in gossip networks
We introduce models of gossip based communication networks in which each node
is simultaneously a sensor, a relay and a user of information. We model the
status of ages of information between nodes as a discrete time Markov chain. In
this setting a gossip transmission policy is a decision made at each node
regarding what type of information to relay at any given time (if any). When
transmission policies are based on random decisions, we are able to analyze the
age of information in certain illustrative structured examples either by means
of an explicit analysis, an algorithm or asymptotic approximations. Our key
contribution is presenting this class of models.Comment: 15 pages, 8 figure
Control Aware Radio Resource Allocation in Low Latency Wireless Control Systems
We consider the problem of allocating radio resources over wireless
communication links to control a series of independent wireless control
systems. Low-latency transmissions are necessary in enabling time-sensitive
control systems to operate over wireless links with high reliability. Achieving
fast data rates over wireless links thus comes at the cost of reliability in
the form of high packet error rates compared to wired links due to channel
noise and interference. However, the effect of the communication link errors on
the control system performance depends dynamically on the control system state.
We propose a novel control-communication co-design approach to the low-latency
resource allocation problem. We incorporate control and channel state
information to make scheduling decisions over time on frequency, bandwidth and
data rates across the next-generation Wi-Fi based wireless communication links
that close the control loops. Control systems that are closer to instability or
further from a desired range in a given control cycle are given higher packet
delivery rate targets to meet. Rather than a simple priority ranking, we derive
precise packet error rate targets for each system needed to satisfy stability
targets and make scheduling decisions to meet such targets while reducing total
transmission time. The resulting Control-Aware Low Latency Scheduling (CALLS)
method is tested in numerous simulation experiments that demonstrate its
effectiveness in meeting control-based goals under tight latency constraints
relative to control-agnostic scheduling
- …