30 research outputs found
Flow Smoothing and Denoising: Graph Signal Processing in the Edge-Space
This paper focuses on devising graph signal processing tools for the
treatment of data defined on the edges of a graph. We first show that
conventional tools from graph signal processing may not be suitable for the
analysis of such signals. More specifically, we discuss how the underlying
notion of a `smooth signal' inherited from (the typically considered variants
of) the graph Laplacian are not suitable when dealing with edge signals that
encode a notion of flow. To overcome this limitation we introduce a class of
filters based on the Edge-Laplacian, a special case of the Hodge-Laplacian for
simplicial complexes of order one. We demonstrate how this Edge-Laplacian leads
to low-pass filters that enforce (approximate) flow-conservation in the
processed signals. Moreover, we show how these new filters can be combined with
more classical Laplacian-based processing methods on the line-graph. Finally,
we illustrate the developed tools by denoising synthetic traffic flows on the
London street network.Comment: 5 pages, 2 figur
Time-varying Signals Recovery via Graph Neural Networks
The recovery of time-varying graph signals is a fundamental problem with
numerous applications in sensor networks and forecasting in time series.
Effectively capturing the spatio-temporal information in these signals is
essential for the downstream tasks. Previous studies have used the smoothness
of the temporal differences of such graph signals as an initial assumption.
Nevertheless, this smoothness assumption could result in a degradation of
performance in the corresponding application when the prior does not hold. In
this work, we relax the requirement of this hypothesis by including a learning
module. We propose a Time Graph Neural Network (TimeGNN) for the recovery of
time-varying graph signals. Our algorithm uses an encoder-decoder architecture
with a specialized loss composed of a mean squared error function and a Sobolev
smoothness operator.TimeGNN shows competitive performance against previous
methods in real datasets.Comment: Published in IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP) 2023, Greec
Sampling of graph signals via randomized local aggregations
Sampling of signals defined over the nodes of a graph is one of the crucial
problems in graph signal processing. While in classical signal processing
sampling is a well defined operation, when we consider a graph signal many new
challenges arise and defining an efficient sampling strategy is not
straightforward. Recently, several works have addressed this problem. The most
common techniques select a subset of nodes to reconstruct the entire signal.
However, such methods often require the knowledge of the signal support and the
computation of the sparsity basis before sampling. Instead, in this paper we
propose a new approach to this issue. We introduce a novel technique that
combines localized sampling with compressed sensing. We first choose a subset
of nodes and then, for each node of the subset, we compute random linear
combinations of signal coefficients localized at the node itself and its
neighborhood. The proposed method provides theoretical guarantees in terms of
reconstruction and stability to noise for any graph and any orthonormal basis,
even when the support is not known.Comment: IEEE Transactions on Signal and Information Processing over Networks,
201
Adaptive Graph Signal Processing: Algorithms and Optimal Sampling Strategies
The goal of this paper is to propose novel strategies for adaptive learning
of signals defined over graphs, which are observed over a (randomly
time-varying) subset of vertices. We recast two classical adaptive algorithms
in the graph signal processing framework, namely, the least mean squares (LMS)
and the recursive least squares (RLS) adaptive estimation strategies. For both
methods, a detailed mean-square analysis illustrates the effect of random
sampling on the adaptive reconstruction capability and the steady-state
performance. Then, several probabilistic sampling strategies are proposed to
design the sampling probability at each node in the graph, with the aim of
optimizing the tradeoff between steady-state performance, graph sampling rate,
and convergence rate of the adaptive algorithms. Finally, a distributed RLS
strategy is derived and is shown to be convergent to its centralized
counterpart. Numerical simulations carried out over both synthetic and real
data illustrate the good performance of the proposed sampling and
reconstruction strategies for (possibly distributed) adaptive learning of
signals defined over graphs.Comment: Submitted to IEEE Transactions on Signal Processing, September 201