106,216 research outputs found
Diffusion Adaptation Strategies for Distributed Optimization and Learning over Networks
We propose an adaptive diffusion mechanism to optimize a global cost function
in a distributed manner over a network of nodes. The cost function is assumed
to consist of a collection of individual components. Diffusion adaptation
allows the nodes to cooperate and diffuse information in real-time; it also
helps alleviate the effects of stochastic gradient noise and measurement noise
through a continuous learning process. We analyze the mean-square-error
performance of the algorithm in some detail, including its transient and
steady-state behavior. We also apply the diffusion algorithm to two problems:
distributed estimation with sparse parameters and distributed localization.
Compared to well-studied incremental methods, diffusion methods do not require
the use of a cyclic path over the nodes and are robust to node and link
failure. Diffusion methods also endow networks with adaptation abilities that
enable the individual nodes to continue learning even when the cost function
changes with time. Examples involving such dynamic cost functions with moving
targets are common in the context of biological networks.Comment: 34 pages, 6 figures, to appear in IEEE Transactions on Signal
Processing, 201
Incremental Predictive Process Monitoring: How to Deal with the Variability of Real Environments
A characteristic of existing predictive process monitoring techniques is to
first construct a predictive model based on past process executions, and then
use it to predict the future of new ongoing cases, without the possibility of
updating it with new cases when they complete their execution. This can make
predictive process monitoring too rigid to deal with the variability of
processes working in real environments that continuously evolve and/or exhibit
new variant behaviors over time. As a solution to this problem, we propose the
use of algorithms that allow the incremental construction of the predictive
model. These incremental learning algorithms update the model whenever new
cases become available so that the predictive model evolves over time to fit
the current circumstances. The algorithms have been implemented using different
case encoding strategies and evaluated on a number of real and synthetic
datasets. The results provide a first evidence of the potential of incremental
learning strategies for predicting process monitoring in real environments, and
of the impact of different case encoding strategies in this setting
Distributed Diffusion-based LMS for Node-Specific Parameter Estimation over Adaptive Networks
A distributed adaptive algorithm is proposed to solve a node-specific
parameter estimation problem where nodes are interested in estimating
parameters of local interest and parameters of global interest to the whole
network. To address the different node-specific parameter estimation problems,
this novel algorithm relies on a diffusion-based implementation of different
Least Mean Squares (LMS) algorithms, each associated with the estimation of a
specific set of local or global parameters. Although all the different LMS
algorithms are coupled, the diffusion-based implementation of each LMS
algorithm is exclusively undertaken by the nodes of the network interested in a
specific set of local or global parameters. To illustrate the effectiveness of
the proposed technique we provide simulation results in the context of
cooperative spectrum sensing in cognitive radio networks.Comment: 5 pages, 2 figures, Published in Proc. IEEE ICASSP, Florence, Italy,
May 201
- …