6,481 research outputs found

    CayleyNets: Graph Convolutional Neural Networks with Complex Rational Spectral Filters

    Full text link
    The rise of graph-structured data such as social networks, regulatory networks, citation graphs, and functional brain networks, in combination with resounding success of deep learning in various applications, has brought the interest in generalizing deep learning models to non-Euclidean domains. In this paper, we introduce a new spectral domain convolutional architecture for deep learning on graphs. The core ingredient of our model is a new class of parametric rational complex functions (Cayley polynomials) allowing to efficiently compute spectral filters on graphs that specialize on frequency bands of interest. Our model generates rich spectral filters that are localized in space, scales linearly with the size of the input data for sparsely-connected graphs, and can handle different constructions of Laplacian operators. Extensive experimental results show the superior performance of our approach, in comparison to other spectral domain convolutional architectures, on spectral image classification, community detection, vertex classification and matrix completion tasks

    The geometric mean of two matrices from a computational viewpoint

    Full text link
    The geometric mean of two matrices is considered and analyzed from a computational viewpoint. Some useful theoretical properties are derived and an analysis of the conditioning is performed. Several numerical algorithms based on different properties and representation of the geometric mean are discussed and analyzed and it is shown that most of them can be classified in terms of the rational approximations of the inverse square root functions. A review of the relevant applications is given

    On the spectral distribution of kernel matrices related to\ud radial basis functions

    Get PDF
    This paper focuses on the spectral distribution of kernel matrices related to radial basis functions. The asymptotic behaviour of eigenvalues of kernel matrices related to radial basis functions with different smoothness are studied. These results are obtained by estimated the coefficients of an orthogonal expansion of the underlying kernel function. Beside many other results, we prove that there are exactly (k+d−1/d-1) eigenvalues in the same order for analytic separable kernel functions like the Gaussian in Rd. This gives theoretical support for how to choose the diagonal scaling matrix in the RBF-QR method (Fornberg et al, SIAM J. Sci. Comput. (33), 2011) which can stably compute Gaussian radial basis function interpolants
    • …
    corecore