1,531 research outputs found
Distributed Local Linear Parameter Estimation using Gaussian SPAWN
We consider the problem of estimating local sensor parameters, where the
local parameters and sensor observations are related through linear stochastic
models. Sensors exchange messages and cooperate with each other to estimate
their own local parameters iteratively. We study the Gaussian Sum-Product
Algorithm over a Wireless Network (gSPAWN) procedure, which is based on belief
propagation, but uses fixed size broadcast messages at each sensor instead.
Compared with the popular diffusion strategies for performing network parameter
estimation, whose communication cost at each sensor increases with increasing
network density, the gSPAWN algorithm allows sensors to broadcast a message
whose size does not depend on the network size or density, making it more
suitable for applications in wireless sensor networks. We show that the gSPAWN
algorithm converges in mean and has mean-square stability under some technical
sufficient conditions, and we describe an application of the gSPAWN algorithm
to a network localization problem in non-line-of-sight environments. Numerical
results suggest that gSPAWN converges much faster in general than the diffusion
method, and has lower communication costs, with comparable root mean square
errors
Diffusion Adaptation Strategies for Distributed Estimation over Gaussian Markov Random Fields
The aim of this paper is to propose diffusion strategies for distributed
estimation over adaptive networks, assuming the presence of spatially
correlated measurements distributed according to a Gaussian Markov random field
(GMRF) model. The proposed methods incorporate prior information about the
statistical dependency among observations, while at the same time processing
data in real-time and in a fully decentralized manner. A detailed mean-square
analysis is carried out in order to prove stability and evaluate the
steady-state performance of the proposed strategies. Finally, we also
illustrate how the proposed techniques can be easily extended in order to
incorporate thresholding operators for sparsity recovery applications.
Numerical results show the potential advantages of using such techniques for
distributed learning in adaptive networks deployed over GMRF.Comment: Submitted to IEEE Transactions on Signal Processing. arXiv admin
note: text overlap with arXiv:1206.309
Diffusion Adaptation over Networks under Imperfect Information Exchange and Non-stationary Data
Adaptive networks rely on in-network and collaborative processing among
distributed agents to deliver enhanced performance in estimation and inference
tasks. Information is exchanged among the nodes, usually over noisy links. The
combination weights that are used by the nodes to fuse information from their
neighbors play a critical role in influencing the adaptation and tracking
abilities of the network. This paper first investigates the mean-square
performance of general adaptive diffusion algorithms in the presence of various
sources of imperfect information exchanges, quantization errors, and model
non-stationarities. Among other results, the analysis reveals that link noise
over the regression data modifies the dynamics of the network evolution in a
distinct way, and leads to biased estimates in steady-state. The analysis also
reveals how the network mean-square performance is dependent on the combination
weights. We use these observations to show how the combination weights can be
optimized and adapted. Simulation results illustrate the theoretical findings
and match well with theory.Comment: 36 pages, 7 figures, to appear in IEEE Transactions on Signal
Processing, June 201
Multitask Diffusion Adaptation over Networks
Adaptive networks are suitable for decentralized inference tasks, e.g., to
monitor complex natural phenomena. Recent research works have intensively
studied distributed optimization problems in the case where the nodes have to
estimate a single optimum parameter vector collaboratively. However, there are
many important applications that are multitask-oriented in the sense that there
are multiple optimum parameter vectors to be inferred simultaneously, in a
collaborative manner, over the area covered by the network. In this paper, we
employ diffusion strategies to develop distributed algorithms that address
multitask problems by minimizing an appropriate mean-square error criterion
with -regularization. The stability and convergence of the algorithm in
the mean and in the mean-square sense is analyzed. Simulations are conducted to
verify the theoretical findings, and to illustrate how the distributed strategy
can be used in several useful applications related to spectral sensing, target
localization, and hyperspectral data unmixing.Comment: 29 pages, 11 figures, submitted for publicatio
Proximal Multitask Learning over Networks with Sparsity-inducing Coregularization
In this work, we consider multitask learning problems where clusters of nodes
are interested in estimating their own parameter vector. Cooperation among
clusters is beneficial when the optimal models of adjacent clusters have a good
number of similar entries. We propose a fully distributed algorithm for solving
this problem. The approach relies on minimizing a global mean-square error
criterion regularized by non-differentiable terms to promote cooperation among
neighboring clusters. A general diffusion forward-backward splitting strategy
is introduced. Then, it is specialized to the case of sparsity promoting
regularizers. A closed-form expression for the proximal operator of a weighted
sum of -norms is derived to achieve higher efficiency. We also provide
conditions on the step-sizes that ensure convergence of the algorithm in the
mean and mean-square error sense. Simulations are conducted to illustrate the
effectiveness of the strategy
Distributed Adaptive Learning of Graph Signals
The aim of this paper is to propose distributed strategies for adaptive
learning of signals defined over graphs. Assuming the graph signal to be
bandlimited, the method enables distributed reconstruction, with guaranteed
performance in terms of mean-square error, and tracking from a limited number
of sampled observations taken from a subset of vertices. A detailed mean square
analysis is carried out and illustrates the role played by the sampling
strategy on the performance of the proposed method. Finally, some useful
strategies for distributed selection of the sampling set are provided. Several
numerical results validate our theoretical findings, and illustrate the
performance of the proposed method for distributed adaptive learning of signals
defined over graphs.Comment: To appear in IEEE Transactions on Signal Processing, 201
- …