12,194 research outputs found
Online Learning Algorithm for Time Series Forecasting Suitable for Low Cost Wireless Sensor Networks Nodes
Time series forecasting is an important predictive methodology which can be
applied to a wide range of problems. Particularly, forecasting the indoor
temperature permits an improved utilization of the HVAC (Heating, Ventilating
and Air Conditioning) systems in a home and thus a better energy efficiency.
With such purpose the paper describes how to implement an Artificial Neural
Network (ANN) algorithm in a low cost system-on-chip to develop an autonomous
intelligent wireless sensor network. The present paper uses a Wireless Sensor
Networks (WSN) to monitor and forecast the indoor temperature in a smart home,
based on low resources and cost microcontroller technology as the 8051MCU. An
on-line learning approach, based on Back-Propagation (BP) algorithm for ANNs,
has been developed for real-time time series learning. It performs the model
training with every new data that arrive to the system, without saving enormous
quantities of data to create a historical database as usual, i.e., without
previous knowledge. Consequently to validate the approach a simulation study
through a Bayesian baseline model have been tested in order to compare with a
database of a real application aiming to see the performance and accuracy. The
core of the paper is a new algorithm, based on the BP one, which has been
described in detail, and the challenge was how to implement a computational
demanding algorithm in a simple architecture with very few hardware resources.Comment: 28 pages, Published 21 April 2015 at MDPI's journal "Sensors
Matrix-Monotonic Optimization for MIMO Systems
For MIMO systems, due to the deployment of multiple antennas at both the
transmitter and the receiver, the design variables e.g., precoders, equalizers,
training sequences, etc. are usually matrices. It is well known that matrix
operations are usually more complicated compared to their vector counterparts.
In order to overcome the high complexity resulting from matrix variables, in
this paper we investigate a class of elegant multi-objective optimization
problems, namely matrix-monotonic optimization problems (MMOPs). In our work,
various representative MIMO optimization problems are unified into a framework
of matrix-monotonic optimization, which includes linear transceiver design,
nonlinear transceiver design, training sequence design, radar waveform
optimization, the corresponding robust design and so on as its special cases.
Then exploiting the framework of matrix-monotonic optimization the optimal
structures of the considered matrix variables can be derived first. Based on
the optimal structure, the matrix-variate optimization problems can be greatly
simplified into the ones with only vector variables. In particular, the
dimension of the new vector variable is equal to the minimum number of columns
and rows of the original matrix variable. Finally, we also extend our work to
some more general cases with multiple matrix variables.Comment: 37 Pages, 5 figures, IEEE Transactions on Signal Processing, Final
Versio
A cell outage management framework for dense heterogeneous networks
In this paper, we present a novel cell outage management (COM) framework for heterogeneous networks with split control and data planes-a candidate architecture for meeting future capacity, quality-of-service, and energy efficiency demands. In such an architecture, the control and data functionalities are not necessarily handled by the same node. The control base stations (BSs) manage the transmission of control information and user equipment (UE) mobility, whereas the data BSs handle UE data. An implication of this split architecture is that an outage to a BS in one plane has to be compensated by other BSs in the same plane. Our COM framework addresses this challenge by incorporating two distinct cell outage detection (COD) algorithms to cope with the idiosyncrasies of both data and control planes. The COD algorithm for control cells leverages the relatively larger number of UEs in the control cell to gather large-scale minimization-of-drive-test report data and detects an outage by applying machine learning and anomaly detection techniques. To improve outage detection accuracy, we also investigate and compare the performance of two anomaly-detecting algorithms, i.e., k-nearest-neighbor- and local-outlier-factor-based anomaly detectors, within the control COD. On the other hand, for data cell COD, we propose a heuristic Grey-prediction-based approach, which can work with the small number of UE in the data cell, by exploiting the fact that the control BS manages UE-data BS connectivity and by receiving a periodic update of the received signal reference power statistic between the UEs and data BSs in its coverage. The detection accuracy of the heuristic data COD algorithm is further improved by exploiting the Fourier series of the residual error that is inherent to a Grey prediction model. Our COM framework integrates these two COD algorithms with a cell outage compensation (COC) algorithm that can be applied to both planes. Our COC solution utilizes an actor-critic-based reinforcement learning algorithm, which optimizes the capacity and coverage of the identified outage zone in a plane, by adjusting the antenna gain and transmission power of the surrounding BSs in that plane. The simulation results show that the proposed framework can detect both data and control cell outage and compensate for the detected outage in a reliable manner
In-Network Distributed Solar Current Prediction
Long-term sensor network deployments demand careful power management. While
managing power requires understanding the amount of energy harvestable from the
local environment, current solar prediction methods rely only on recent local
history, which makes them susceptible to high variability. In this paper, we
present a model and algorithms for distributed solar current prediction, based
on multiple linear regression to predict future solar current based on local,
in-situ climatic and solar measurements. These algorithms leverage spatial
information from neighbors and adapt to the changing local conditions not
captured by global climatic information. We implement these algorithms on our
Fleck platform and run a 7-week-long experiment validating our work. In
analyzing our results from this experiment, we determined that computing our
model requires an increased energy expenditure of 4.5mJ over simpler models (on
the order of 10^{-7}% of the harvested energy) to gain a prediction improvement
of 39.7%.Comment: 28 pages, accepted at TOSN and awaiting publicatio
Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks
Future wireless networks have a substantial potential in terms of supporting
a broad range of complex compelling applications both in military and civilian
fields, where the users are able to enjoy high-rate, low-latency, low-cost and
reliable information services. Achieving this ambitious goal requires new radio
techniques for adaptive learning and intelligent decision making because of the
complex heterogeneous nature of the network structures and wireless services.
Machine learning (ML) algorithms have great success in supporting big data
analytics, efficient parameter estimation and interactive decision making.
Hence, in this article, we review the thirty-year history of ML by elaborating
on supervised learning, unsupervised learning, reinforcement learning and deep
learning. Furthermore, we investigate their employment in the compelling
applications of wireless networks, including heterogeneous networks (HetNets),
cognitive radios (CR), Internet of things (IoT), machine to machine networks
(M2M), and so on. This article aims for assisting the readers in clarifying the
motivation and methodology of the various ML algorithms, so as to invoke them
for hitherto unexplored services as well as scenarios of future wireless
networks.Comment: 46 pages, 22 fig
- …