15,424 research outputs found

    Propagation Kernels

    Full text link
    We introduce propagation kernels, a general graph-kernel framework for efficiently measuring the similarity of structured data. Propagation kernels are based on monitoring how information spreads through a set of given graphs. They leverage early-stage distributions from propagation schemes such as random walks to capture structural information encoded in node labels, attributes, and edge information. This has two benefits. First, off-the-shelf propagation schemes can be used to naturally construct kernels for many graph types, including labeled, partially labeled, unlabeled, directed, and attributed graphs. Second, by leveraging existing efficient and informative propagation schemes, propagation kernels can be considerably faster than state-of-the-art approaches without sacrificing predictive performance. We will also show that if the graphs at hand have a regular structure, for instance when modeling image or video data, one can exploit this regularity to scale the kernel computation to large databases of graphs with thousands of nodes. We support our contributions by exhaustive experiments on a number of real-world graphs from a variety of application domains

    A Survey on Graph Kernels

    Get PDF
    Graph kernels have become an established and widely-used technique for solving classification tasks on graphs. This survey gives a comprehensive overview of techniques for kernel-based graph classification developed in the past 15 years. We describe and categorize graph kernels based on properties inherent to their design, such as the nature of their extracted graph features, their method of computation and their applicability to problems in practice. In an extensive experimental evaluation, we study the classification accuracy of a large suite of graph kernels on established benchmarks as well as new datasets. We compare the performance of popular kernels with several baseline methods and study the effect of applying a Gaussian RBF kernel to the metric induced by a graph kernel. In doing so, we find that simple baselines become competitive after this transformation on some datasets. Moreover, we study the extent to which existing graph kernels agree in their predictions (and prediction errors) and obtain a data-driven categorization of kernels as result. Finally, based on our experimental results, we derive a practitioner's guide to kernel-based graph classification

    DeepSphere: Efficient spherical Convolutional Neural Network with HEALPix sampling for cosmological applications

    Full text link
    Convolutional Neural Networks (CNNs) are a cornerstone of the Deep Learning toolbox and have led to many breakthroughs in Artificial Intelligence. These networks have mostly been developed for regular Euclidean domains such as those supporting images, audio, or video. Because of their success, CNN-based methods are becoming increasingly popular in Cosmology. Cosmological data often comes as spherical maps, which make the use of the traditional CNNs more complicated. The commonly used pixelization scheme for spherical maps is the Hierarchical Equal Area isoLatitude Pixelisation (HEALPix). We present a spherical CNN for analysis of full and partial HEALPix maps, which we call DeepSphere. The spherical CNN is constructed by representing the sphere as a graph. Graphs are versatile data structures that can act as a discrete representation of a continuous manifold. Using the graph-based representation, we define many of the standard CNN operations, such as convolution and pooling. With filters restricted to being radial, our convolutions are equivariant to rotation on the sphere, and DeepSphere can be made invariant or equivariant to rotation. This way, DeepSphere is a special case of a graph CNN, tailored to the HEALPix sampling of the sphere. This approach is computationally more efficient than using spherical harmonics to perform convolutions. We demonstrate the method on a classification problem of weak lensing mass maps from two cosmological models and compare the performance of the CNN with that of two baseline classifiers. The results show that the performance of DeepSphere is always superior or equal to both of these baselines. For high noise levels and for data covering only a smaller fraction of the sphere, DeepSphere achieves typically 10% better classification accuracy than those baselines. Finally, we show how learned filters can be visualized to introspect the neural network.Comment: arXiv admin note: text overlap with arXiv:astro-ph/0409513 by other author

    The Evolution of Correlation Functions in the Zel'dovich Approximation and its Implications for the Validity of Perturbation Theory

    Get PDF
    We investigate whether it is possible to study perturbatively the transition in cosmological clustering between a single streamed flow to a multi streamed flow. We do this by considereing a system whose dynamics is governed by the Zel'dovich approximation (ZA) and calculating the evolution of the two point correlation function using two methods: 1.Distribution functions 2.Hydrodynamic equations without pressure and vorticity. The latter method breaks down once multistreaming occurs whereas the former does not. We find that the two methods give the same results to all orders in a perturbative expansion of the two point correlation function. We thus conclude that we cannot study the transition from a single stream flow to a multi-stream flow in a perturbative expansion. We expect this conclusion to hold even if we use the full gravitational dynamics (GD) instead of ZA. We use ZA to look at the evolution of the two point correlation function at large spatial separations and we find that until the onset of multi-streaming the evolution can be described by a diffusion process where the linear evolution at large scales gets modified by the rearrangement of matter on small scales. We compare these results with the lowest order nonlinear results from GD. We find that the difference is only in the numerical value of the diffusion coefficient and we interpret this physically. We also use ZA to study the induced three point correlation function. At the lowest order we find that, as in the case of GD, the three point correlation does not necessarily have the hierarchical form. We also find that at large separations the effect of the higher order terms for the three point correlatin function is very similar to that for the the two point correlation and in this case too the evolution can be be described in terms ofComment: 28 pages including 6 figures, Latex, Aastex macros, Accepted in Astrophysical Journa

    Geometric deep learning

    Get PDF
    The goal of these course notes is to describe the main mathematical ideas behind geometric deep learning and to provide implementation details for several applications in shape analysis and synthesis, computer vision and computer graphics. The text in the course materials is primarily based on previously published work. With these notes we gather and provide a clear picture of the key concepts and techniques that fall under the umbrella of geometric deep learning, and illustrate the applications they enable. We also aim to provide practical implementation details for the methods presented in these works, as well as suggest further readings and extensions of these ideas

    The Feynman graph representation of convolution semigroups and its applications to L\'{e}vy statistics

    Full text link
    We consider the Cauchy problem for a pseudo-differential operator which has a translation-invariant and analytic symbol. For a certain set of initial conditions, a formal solution is obtained by a perturbative expansion. The series so obtained can be re-expressed in terms of generalized Feynman graphs and Feynman rules. The logarithm of the solution can then be represented by a series containing only the connected Feynman graphs. Under some conditions, it is shown that the formal solution uniquely determines the real solution by means of Borel transforms. The formalism is then applied to probabilistic L\'{e}vy distributions. Here, the Gaussian part of such a distribution is re-interpreted as a initial condition and a large diffusion expansion for L\'{e}vy densities is obtained. It is outlined how this expansion can be used in statistical problems that involve L\'{e}vy distributions.Comment: Published in at http://dx.doi.org/10.3150/07-BEJ106 the Bernoulli (http://isi.cbs.nl/bernoulli/) by the International Statistical Institute/Bernoulli Society (http://isi.cbs.nl/BS/bshome.htm

    Nonlocal, noncommutative diagrammatics and the linked cluster Theorems

    Get PDF
    Recent developments in quantum chemistry, perturbative quantum field theory, statistical physics or stochastic differential equations require the introduction of new families of Feynman-type diagrams. These new families arise in various ways. In some generalizations of the classical diagrams, the notion of Feynman propagator is extended to generalized propagators connecting more than two vertices of the graphs. In some others (introduced in the present article), the diagrams, associated to noncommuting product of operators inherit from the noncommutativity of the products extra graphical properties. The purpose of the present article is to introduce a general way of dealing with such diagrams. We prove in particular a "universal" linked cluster theorem and introduce, in the process, a Feynman-type "diagrammatics" that allows to handle simultaneously nonlocal (Coulomb-type) interactions, the generalized diagrams arising from the study of interacting systems (such as the ones where the ground state is not the vacuum but e.g. a vacuum perturbed by a magnetic or electric field, by impurities...) or Wightman fields (that is, expectation values of products of interacting fields). Our diagrammatics seems to be the first attempt to encode in a unified algebraic framework such a wide variety of situations. In the process, we promote two ideas. First, Feynman-type diagrammatics belong mathematically to the theory of linear forms on combinatorial Hopf algebras. Second, linked cluster-type theorems rely ultimately on M\"obius inversion on the partition lattice. The two theories should therefore be introduced and presented accordingl
    corecore