1,726 research outputs found

    Geometric deep learning: going beyond Euclidean data

    Get PDF
    Many scientific fields study data with an underlying structure that is a non-Euclidean space. Some examples include social networks in computational social sciences, sensor networks in communications, functional networks in brain imaging, regulatory networks in genetics, and meshed surfaces in computer graphics. In many applications, such geometric data are large and complex (in the case of social networks, on the scale of billions), and are natural targets for machine learning techniques. In particular, we would like to use deep neural networks, which have recently proven to be powerful tools for a broad range of problems from computer vision, natural language processing, and audio analysis. However, these tools have been most successful on data with an underlying Euclidean or grid-like structure, and in cases where the invariances of these structures are built into networks used to model them. Geometric deep learning is an umbrella term for emerging techniques attempting to generalize (structured) deep neural models to non-Euclidean domains such as graphs and manifolds. The purpose of this paper is to overview different examples of geometric deep learning problems and present available solutions, key difficulties, applications, and future research directions in this nascent field

    Robust and efficient Fourier-Mellin transform approximations for invariant grey-level image description and reconstruction

    No full text
    International audienceThis paper addresses the gray-level image representation ability of the Fourier-Mellin Transform (FMT) for pattern recognition, reconstruction and image database retrieval. The main practical di±culty of the FMT lies in the accuracy and e±ciency of its numerical approximation and we propose three estimations of its analytical extension. Comparison of these approximations is performed from discrete and ¯nite-extent sets of Fourier- Mellin harmonics by means of experiments in: (i) image reconstruction via both visual inspection and the computation of a reconstruction error; and (ii) pattern recognition and discrimination by using a complete and convergent set of features invariant under planar similarities. Experimental results on real gray-level images show that it is possible to recover an image to within a speci¯ed degree of accuracy and to classify objects reliably even when a large set of descriptors is used. Finally, an example will be given, illustrating both theoretical and numerical results in the context of content-based image retrieval

    A new approach to observational cosmology using the scattering transform

    Get PDF
    Parameter estimation with non-Gaussian stochastic fields is a common challenge in astrophysics and cosmology. In this paper we advocate performing this task using the scattering transform, a statistical tool rooted in the mathematical properties of convolutional neural nets. This estimator can characterize a complex field without explicitly computing higher-order statistics, thus avoiding the high variance and dimensionality problems. It generates a compact set of coefficients which can be used as robust summary statistics for non-Gaussian information. It is especially suited for fields presenting localized structures and hierarchical clustering, such as the cosmological density field. To demonstrate its power, we apply this estimator to the cosmological parameter inference problem in the context of weak lensing. Using simulated convergence maps with realistic noise, the scattering transform outperforms the power spectrum and peak counts, and is on par with the state-of-the-art CNN. It retains the advantages of traditional statistical descriptors (it does not require any training nor tuning), has provable stability properties, allows to check for systematics, and importantly, the scattering coefficients are interpretable. It is a powerful and attractive estimator for observational cosmology and, in general, the study of physically-motivated fields.Comment: 15 pages, 7 figures; comments welcom

    Interpretable Transformations with Encoder-Decoder Networks

    Full text link
    Deep feature spaces have the capacity to encode complex transformations of their input data. However, understanding the relative feature-space relationship between two transformed encoded images is difficult. For instance, what is the relative feature space relationship between two rotated images? What is decoded when we interpolate in feature space? Ideally, we want to disentangle confounding factors, such as pose, appearance, and illumination, from object identity. Disentangling these is difficult because they interact in very nonlinear ways. We propose a simple method to construct a deep feature space, with explicitly disentangled representations of several known transformations. A person or algorithm can then manipulate the disentangled representation, for example, to re-render an image with explicit control over parameterized degrees of freedom. The feature space is constructed using a transforming encoder-decoder network with a custom feature transform layer, acting on the hidden representations. We demonstrate the advantages of explicit disentangling on a variety of datasets and transformations, and as an aid for traditional tasks, such as classification.Comment: Accepted at ICCV 201

    A Stable Multi-Scale Kernel for Topological Machine Learning

    Full text link
    Topological data analysis offers a rich source of valuable information to study vision problems. Yet, so far we lack a theoretically sound connection to popular kernel-based learning techniques, such as kernel SVMs or kernel PCA. In this work, we establish such a connection by designing a multi-scale kernel for persistence diagrams, a stable summary representation of topological features in data. We show that this kernel is positive definite and prove its stability with respect to the 1-Wasserstein distance. Experiments on two benchmark datasets for 3D shape classification/retrieval and texture recognition show considerable performance gains of the proposed method compared to an alternative approach that is based on the recently introduced persistence landscapes
    corecore