6 research outputs found
Extension of Recurrent Kernels to different Reservoir Computing topologies
Reservoir Computing (RC) has become popular in recent years due to its fast
and efficient computational capabilities. Standard RC has been shown to be
equivalent in the asymptotic limit to Recurrent Kernels, which helps in
analyzing its expressive power. However, many well-established RC paradigms,
such as Leaky RC, Sparse RC, and Deep RC, are yet to be analyzed in such a way.
This study aims to fill this gap by providing an empirical analysis of the
equivalence of specific RC architectures with their corresponding Recurrent
Kernel formulation. We conduct a convergence study by varying the activation
function implemented in each architecture. Our study also sheds light on the
role of sparse connections in RC architectures and propose an optimal sparsity
level that depends on the reservoir size. Furthermore, our systematic analysis
shows that in Deep RC models, convergence is better achieved with successive
reservoirs of decreasing sizes.Comment: 8 page
On the approximation capability of GNNs in node classification/regression tasks
Graph neural networks (GNNs) are a broad class of connectionist models for graph processing. Recent studies have shown that GNNs can approximate any function on graphs, modulo the equivalence relation on graphs defined by theWeisfeiler-Lehman (WL) test. However, these results suffer from some limitations, both because they were derived using the Stone-Weierstrass theorem — which is existential in nature — and because they assume that the target function to be approximated must be continuous. Furthermore, all current results are dedicated to graph classification/regression tasks, where the GNN must produce a single output for the whole graph, while also node classification/regression problems, in which an output is returned for each node, are very common. In this paper, we propose an alternative way to demonstrate the approximation
capability of GNNs that overcomes these limitations. Indeed, we show that GNNs are universal approximators in probability for node classification/regression tasks, as they can approximate any measurable function that satisfies the 1-WL-equivalence on nodes. The proposed theoretical framework allows the approximation of generic discontinuous target functions and also suggests the GNN architecture that can reach a desired approximation. In addition, we provide a bound on the number of the GNN layers required to achieve the desired degree of approximation, namely 2r − 1, where r is the maximum number of nodes for the graphs in the domain
Weisfeiler--Lehman goes Dynamic: An Analysis of the Expressive Power of Graph Neural Networks for Attributed and Dynamic Graphs
Graph Neural Networks (GNNs) are a large class of relational models for graph
processing. Recent theoretical studies on the expressive power of GNNs have
focused on two issues. On the one hand, it has been proven that GNNs are as
powerful as the Weisfeiler-Lehman test (1-WL) in their ability to distinguish
graphs. Moreover, it has been shown that the equivalence enforced by 1-WL
equals unfolding equivalence. On the other hand, GNNs turned out to be
universal approximators on graphs modulo the constraints enforced by
1-WL/unfolding equivalence. However, these results only apply to Static
Undirected Homogeneous Graphs with node attributes. In contrast, real-life
applications often involve a variety of graph properties, such as, e.g.,
dynamics or node and edge attributes. In this paper, we conduct a theoretical
analysis of the expressive power of GNNs for these two graph types that are
particularly of interest. Dynamic graphs are widely used in modern
applications, and its theoretical analysis requires new approaches. The
attributed type acts as a standard form for all graph types since it has been
shown that all graph types can be transformed without loss to Static Undirected
Homogeneous Graphs with attributes on nodes and edges (SAUHG). The study
considers generic GNN models and proposes appropriate 1-WL tests for those
domains. Then, the results on the expressive power of GNNs are extended by
proving that GNNs have the same capability as the 1-WL test in distinguishing
dynamic and attributed graphs, the 1-WL equivalence equals unfolding
equivalence and that GNNs are universal approximators modulo 1-WL/unfolding
equivalence. Moreover, the proof of the approximation capability holds for
SAUHGs, which include most of those used in practical applications, and it is
constructive in nature allowing to deduce hints on the architecture of GNNs
that can achieve the desired accuracy
Theoretical Properties of Graph Neural Networks
Graph Neural Networks (GNNs) have emerged in recent years as a powerful tool to learn tasks across a wide range of graph domains in a data-driven fashion; based on a message passing mechanism, GNNs have gained increasing popularity due to their intuitive formulation, closely linked with the Weisfeiler-Lehman (WL) test for graph isomorphism, to which they have proven equivalent. In this thesis, we provide a broad overview of two essential properties of GNNs by a theoretical point of view, namely, their approximation power and their generalization capabilities. We show that modern GNNs are universal approximators, given that they are made by a sufficient number of layers, which is tightly linked to the stable node coloring of the 1-WL test. GNNs are shown to be universal approximators also on more complex graph domains, like edge-attributed graphs and dynamic graphs. Generalization capabilities of GNNs are investigated by different perspectives. Bounds on the VC dimension of GNNs are provided with respect to the usual hyperparameters and with respect to the number of colors derived from the 1-WL test. GNNs ability to generalize to unseen data is also explored by a neurocognitive point of view,
determining whether these models are able to learn the so-called identity effects
A unifying point of view on expressive power of GNNs
Graph Neural Networks (GNNs) are a wide class of connectionist models for
graph processing. They perform an iterative message passing operation on each
node and its neighbors, to solve classification/ clustering tasks -- on some
nodes or on the whole graph -- collecting all such messages, regardless of
their order. Despite the differences among the various models belonging to this
class, most of them adopt the same computation scheme, based on a local
aggregation mechanism and, intuitively, the local computation framework is
mainly responsible for the expressive power of GNNs. In this paper, we prove
that the Weisfeiler--Lehman test induces an equivalence relationship on the
graph nodes that exactly corresponds to the unfolding equivalence, defined on
the original GNN model. Therefore, the results on the expressive power of the
original GNNs can be extended to general GNNs which, under mild conditions, can
be proved capable of approximating, in probability and up to any precision, any
function on graphs that respects the unfolding equivalence.Comment: 16 pages, 3 figure
Splines Parameterization of Planar Domains by Physics-Informed Neural Networks
The generation of structured grids on bounded domains is a crucial issue in the development of numerical models for solving differential problems. In particular, the representation of the given computational domain through a regular parameterization allows us to define a univalent mapping, which can be computed as the solution of an elliptic problem, equipped with suitable Dirichlet boundary conditions. In recent years, Physics-Informed Neural Networks (PINNs) have been proved to be a powerful tool to compute the solution of Partial Differential Equations (PDEs) replacing standard numerical models, based on Finite Element Methods and Finite Differences, with deep neural networks; PINNs can be used for predicting the values on simulation grids of different resolutions without the need to be retrained. In this work, we exploit the PINN model in order to solve the PDE associated to the differential problem of the parameterization on both convex and non-convex planar domains, for which the describing PDE is known. The final continuous model is then provided by applying a Hermite type quasi-interpolation operator, which can guarantee the desired smoothness of the sought parameterization. Finally, some numerical examples are presented, which show that the PINNs-based approach is robust. Indeed, the produced mapping does not exhibit folding or self-intersection at the interior of the domain and, also, for highly non convex shapes, despite few faulty points near the boundaries, has better shape-measures, e.g., lower values of the Winslow functional