46,064 research outputs found
Diffusion Adaptation Strategies for Distributed Optimization and Learning over Networks
We propose an adaptive diffusion mechanism to optimize a global cost function
in a distributed manner over a network of nodes. The cost function is assumed
to consist of a collection of individual components. Diffusion adaptation
allows the nodes to cooperate and diffuse information in real-time; it also
helps alleviate the effects of stochastic gradient noise and measurement noise
through a continuous learning process. We analyze the mean-square-error
performance of the algorithm in some detail, including its transient and
steady-state behavior. We also apply the diffusion algorithm to two problems:
distributed estimation with sparse parameters and distributed localization.
Compared to well-studied incremental methods, diffusion methods do not require
the use of a cyclic path over the nodes and are robust to node and link
failure. Diffusion methods also endow networks with adaptation abilities that
enable the individual nodes to continue learning even when the cost function
changes with time. Examples involving such dynamic cost functions with moving
targets are common in the context of biological networks.Comment: 34 pages, 6 figures, to appear in IEEE Transactions on Signal
Processing, 201
A Multitask Diffusion Strategy with Optimized Inter-Cluster Cooperation
We consider a multitask estimation problem where nodes in a network are
divided into several connected clusters, with each cluster performing a
least-mean-squares estimation of a different random parameter vector. Inspired
by the adapt-then-combine diffusion strategy, we propose a multitask diffusion
strategy whose mean stability can be ensured whenever individual nodes are
stable in the mean, regardless of the inter-cluster cooperation weights. In
addition, the proposed strategy is able to achieve an asymptotically unbiased
estimation, when the parameters have same mean. We also develop an
inter-cluster cooperation weights selection scheme that allows each node in the
network to locally optimize its inter-cluster cooperation weights. Numerical
results demonstrate that our approach leads to a lower average steady-state
network mean-square deviation, compared with using weights selected by various
other commonly adopted methods in the literature.Comment: 30 pages, 8 figures, submitted to IEEE Journal of Selected Topics in
Signal Processin
Multi-hop Diffusion LMS for Energy-constrained Distributed Estimation
We propose a multi-hop diffusion strategy for a sensor network to perform
distributed least mean-squares (LMS) estimation under local and network-wide
energy constraints. At each iteration of the strategy, each node can combine
intermediate parameter estimates from nodes other than its physical neighbors
via a multi-hop relay path. We propose a rule to select combination weights for
the multi-hop neighbors, which can balance between the transient and the
steady-state network mean-square deviations (MSDs). We study two classes of
networks: simple networks with a unique transmission path from one node to
another, and arbitrary networks utilizing diffusion consultations over at most
two hops. We propose a method to optimize each node's information neighborhood
subject to local energy budgets and a network-wide energy budget for each
diffusion iteration. This optimization requires the network topology, and the
noise and data variance profiles of each node, and is performed offline before
the diffusion process. In addition, we develop a fully distributed and adaptive
algorithm that approximately optimizes the information neighborhood of each
node with only local energy budget constraints in the case where diffusion
consultations are performed over at most a predefined number of hops. Numerical
results suggest that our proposed multi-hop diffusion strategy achieves the
same steady-state MSD as the existing one-hop adapt-then-combine diffusion
algorithm but with a lower energy budget.Comment: 14 pages, 12 figures. Submitted for publicatio
Adaptation and learning over networks for nonlinear system modeling
In this chapter, we analyze nonlinear filtering problems in distributed
environments, e.g., sensor networks or peer-to-peer protocols. In these
scenarios, the agents in the environment receive measurements in a streaming
fashion, and they are required to estimate a common (nonlinear) model by
alternating local computations and communications with their neighbors. We
focus on the important distinction between single-task problems, where the
underlying model is common to all agents, and multitask problems, where each
agent might converge to a different model due to, e.g., spatial dependencies or
other factors. Currently, most of the literature on distributed learning in the
nonlinear case has focused on the single-task case, which may be a strong
limitation in real-world scenarios. After introducing the problem and reviewing
the existing approaches, we describe a simple kernel-based algorithm tailored
for the multitask case. We evaluate the proposal on a simulated benchmark task,
and we conclude by detailing currently open problems and lines of research.Comment: To be published as a chapter in `Adaptive Learning Methods for
Nonlinear System Modeling', Elsevier Publishing, Eds. D. Comminiello and J.C.
Principe (2018
Multitask Diffusion Adaptation over Networks
Adaptive networks are suitable for decentralized inference tasks, e.g., to
monitor complex natural phenomena. Recent research works have intensively
studied distributed optimization problems in the case where the nodes have to
estimate a single optimum parameter vector collaboratively. However, there are
many important applications that are multitask-oriented in the sense that there
are multiple optimum parameter vectors to be inferred simultaneously, in a
collaborative manner, over the area covered by the network. In this paper, we
employ diffusion strategies to develop distributed algorithms that address
multitask problems by minimizing an appropriate mean-square error criterion
with -regularization. The stability and convergence of the algorithm in
the mean and in the mean-square sense is analyzed. Simulations are conducted to
verify the theoretical findings, and to illustrate how the distributed strategy
can be used in several useful applications related to spectral sensing, target
localization, and hyperspectral data unmixing.Comment: 29 pages, 11 figures, submitted for publicatio
Proximal Multitask Learning over Networks with Sparsity-inducing Coregularization
In this work, we consider multitask learning problems where clusters of nodes
are interested in estimating their own parameter vector. Cooperation among
clusters is beneficial when the optimal models of adjacent clusters have a good
number of similar entries. We propose a fully distributed algorithm for solving
this problem. The approach relies on minimizing a global mean-square error
criterion regularized by non-differentiable terms to promote cooperation among
neighboring clusters. A general diffusion forward-backward splitting strategy
is introduced. Then, it is specialized to the case of sparsity promoting
regularizers. A closed-form expression for the proximal operator of a weighted
sum of -norms is derived to achieve higher efficiency. We also provide
conditions on the step-sizes that ensure convergence of the algorithm in the
mean and mean-square error sense. Simulations are conducted to illustrate the
effectiveness of the strategy
Diffusion LMS for clustered multitask networks
Recent research works on distributed adaptive networks have intensively
studied the case where the nodes estimate a common parameter vector
collaboratively. However, there are many applications that are
multitask-oriented in the sense that there are multiple parameter vectors that
need to be inferred simultaneously. In this paper, we employ diffusion
strategies to develop distributed algorithms that address clustered multitask
problems by minimizing an appropriate mean-square error criterion with
-regularization. Some results on the mean-square stability and
convergence of the algorithm are also provided. Simulations are conducted to
illustrate the theoretical findings.Comment: 5 pages, 6 figures, submitted to ICASSP 201
- …