15,424 research outputs found
Propagation Kernels
We introduce propagation kernels, a general graph-kernel framework for
efficiently measuring the similarity of structured data. Propagation kernels
are based on monitoring how information spreads through a set of given graphs.
They leverage early-stage distributions from propagation schemes such as random
walks to capture structural information encoded in node labels, attributes, and
edge information. This has two benefits. First, off-the-shelf propagation
schemes can be used to naturally construct kernels for many graph types,
including labeled, partially labeled, unlabeled, directed, and attributed
graphs. Second, by leveraging existing efficient and informative propagation
schemes, propagation kernels can be considerably faster than state-of-the-art
approaches without sacrificing predictive performance. We will also show that
if the graphs at hand have a regular structure, for instance when modeling
image or video data, one can exploit this regularity to scale the kernel
computation to large databases of graphs with thousands of nodes. We support
our contributions by exhaustive experiments on a number of real-world graphs
from a variety of application domains
A Survey on Graph Kernels
Graph kernels have become an established and widely-used technique for
solving classification tasks on graphs. This survey gives a comprehensive
overview of techniques for kernel-based graph classification developed in the
past 15 years. We describe and categorize graph kernels based on properties
inherent to their design, such as the nature of their extracted graph features,
their method of computation and their applicability to problems in practice. In
an extensive experimental evaluation, we study the classification accuracy of a
large suite of graph kernels on established benchmarks as well as new datasets.
We compare the performance of popular kernels with several baseline methods and
study the effect of applying a Gaussian RBF kernel to the metric induced by a
graph kernel. In doing so, we find that simple baselines become competitive
after this transformation on some datasets. Moreover, we study the extent to
which existing graph kernels agree in their predictions (and prediction errors)
and obtain a data-driven categorization of kernels as result. Finally, based on
our experimental results, we derive a practitioner's guide to kernel-based
graph classification
DeepSphere: Efficient spherical Convolutional Neural Network with HEALPix sampling for cosmological applications
Convolutional Neural Networks (CNNs) are a cornerstone of the Deep Learning
toolbox and have led to many breakthroughs in Artificial Intelligence. These
networks have mostly been developed for regular Euclidean domains such as those
supporting images, audio, or video. Because of their success, CNN-based methods
are becoming increasingly popular in Cosmology. Cosmological data often comes
as spherical maps, which make the use of the traditional CNNs more complicated.
The commonly used pixelization scheme for spherical maps is the Hierarchical
Equal Area isoLatitude Pixelisation (HEALPix). We present a spherical CNN for
analysis of full and partial HEALPix maps, which we call DeepSphere. The
spherical CNN is constructed by representing the sphere as a graph. Graphs are
versatile data structures that can act as a discrete representation of a
continuous manifold. Using the graph-based representation, we define many of
the standard CNN operations, such as convolution and pooling. With filters
restricted to being radial, our convolutions are equivariant to rotation on the
sphere, and DeepSphere can be made invariant or equivariant to rotation. This
way, DeepSphere is a special case of a graph CNN, tailored to the HEALPix
sampling of the sphere. This approach is computationally more efficient than
using spherical harmonics to perform convolutions. We demonstrate the method on
a classification problem of weak lensing mass maps from two cosmological models
and compare the performance of the CNN with that of two baseline classifiers.
The results show that the performance of DeepSphere is always superior or equal
to both of these baselines. For high noise levels and for data covering only a
smaller fraction of the sphere, DeepSphere achieves typically 10% better
classification accuracy than those baselines. Finally, we show how learned
filters can be visualized to introspect the neural network.Comment: arXiv admin note: text overlap with arXiv:astro-ph/0409513 by other
author
The Evolution of Correlation Functions in the Zel'dovich Approximation and its Implications for the Validity of Perturbation Theory
We investigate whether it is possible to study perturbatively the transition
in cosmological clustering between a single streamed flow to a multi streamed
flow. We do this by considereing a system whose dynamics is governed by the
Zel'dovich approximation (ZA) and calculating the evolution of the two point
correlation function using two methods: 1.Distribution functions 2.Hydrodynamic
equations without pressure and vorticity. The latter method breaks down once
multistreaming occurs whereas the former does not. We find that the two methods
give the same results to all orders in a perturbative expansion of the two
point correlation function. We thus conclude that we cannot study the
transition from a single stream flow to a multi-stream flow in a perturbative
expansion. We expect this conclusion to hold even if we use the full
gravitational dynamics (GD) instead of ZA. We use ZA to look at the evolution
of the two point correlation function at large spatial separations and we find
that until the onset of multi-streaming the evolution can be described by a
diffusion process where the linear evolution at large scales gets modified by
the rearrangement of matter on small scales. We compare these results with the
lowest order nonlinear results from GD. We find that the difference is only in
the numerical value of the diffusion coefficient and we interpret this
physically. We also use ZA to study the induced three point correlation
function. At the lowest order we find that, as in the case of GD, the three
point correlation does not necessarily have the hierarchical form. We also find
that at large separations the effect of the higher order terms for the three
point correlatin function is very similar to that for the the two point
correlation and in this case too the evolution can be be described in terms ofComment: 28 pages including 6 figures, Latex, Aastex macros, Accepted in
Astrophysical Journa
Geometric deep learning
The goal of these course notes is to describe the main mathematical ideas behind geometric deep learning and to provide implementation details for several applications in shape analysis and synthesis, computer vision and computer graphics. The text in the course materials is primarily based on previously published work. With these notes we gather and provide a clear picture of the key concepts and techniques that fall under the umbrella of geometric deep learning, and illustrate the applications they enable. We also aim to provide practical implementation details for the methods presented in these works, as well as suggest further readings and extensions of these ideas
The Feynman graph representation of convolution semigroups and its applications to L\'{e}vy statistics
We consider the Cauchy problem for a pseudo-differential operator which has a
translation-invariant and analytic symbol. For a certain set of initial
conditions, a formal solution is obtained by a perturbative expansion. The
series so obtained can be re-expressed in terms of generalized Feynman graphs
and Feynman rules. The logarithm of the solution can then be represented by a
series containing only the connected Feynman graphs. Under some conditions, it
is shown that the formal solution uniquely determines the real solution by
means of Borel transforms. The formalism is then applied to probabilistic
L\'{e}vy distributions. Here, the Gaussian part of such a distribution is
re-interpreted as a initial condition and a large diffusion expansion for
L\'{e}vy densities is obtained. It is outlined how this expansion can be used
in statistical problems that involve L\'{e}vy distributions.Comment: Published in at http://dx.doi.org/10.3150/07-BEJ106 the Bernoulli
(http://isi.cbs.nl/bernoulli/) by the International Statistical
Institute/Bernoulli Society (http://isi.cbs.nl/BS/bshome.htm
Nonlocal, noncommutative diagrammatics and the linked cluster Theorems
Recent developments in quantum chemistry, perturbative quantum field theory,
statistical physics or stochastic differential equations require the
introduction of new families of Feynman-type diagrams. These new families arise
in various ways. In some generalizations of the classical diagrams, the notion
of Feynman propagator is extended to generalized propagators connecting more
than two vertices of the graphs. In some others (introduced in the present
article), the diagrams, associated to noncommuting product of operators inherit
from the noncommutativity of the products extra graphical properties. The
purpose of the present article is to introduce a general way of dealing with
such diagrams. We prove in particular a "universal" linked cluster theorem and
introduce, in the process, a Feynman-type "diagrammatics" that allows to handle
simultaneously nonlocal (Coulomb-type) interactions, the generalized diagrams
arising from the study of interacting systems (such as the ones where the
ground state is not the vacuum but e.g. a vacuum perturbed by a magnetic or
electric field, by impurities...) or Wightman fields (that is, expectation
values of products of interacting fields). Our diagrammatics seems to be the
first attempt to encode in a unified algebraic framework such a wide variety of
situations. In the process, we promote two ideas. First, Feynman-type
diagrammatics belong mathematically to the theory of linear forms on
combinatorial Hopf algebras. Second, linked cluster-type theorems rely
ultimately on M\"obius inversion on the partition lattice. The two theories
should therefore be introduced and presented accordingl
- …