7,353 research outputs found

    Discriminate-and-Rectify Encoders: Learning from Image Transformation Sets

    Get PDF
    The complexity of a learning task is increased by transformations in the input space that preserve class identity. Visual object recognition for example is affected by changes in viewpoint, scale, illumination or planar transformations. While drastically altering the visual appearance, these changes are orthogonal to recognition and should not be reflected in the representation or feature encoding used for learning. We introduce a framework for weakly supervised learning of image embeddings that are robust to transformations and selective to the class distribution, using sets of transforming examples (orbit sets), deep parametrizations and a novel orbit-based loss. The proposed loss combines a discriminative, contrastive part for orbits with a reconstruction error that learns to rectify orbit transformations. The learned embeddings are evaluated in distance metric-based tasks, such as one-shot classification under geometric transformations, as well as face verification and retrieval under more realistic visual variability. Our results suggest that orbit sets, suitably computed or observed, can be used for efficient, weakly-supervised learning of semantically relevant image embeddings.This material is based upon work supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216

    A new code for orbit analysis and Schwarzschild modelling of triaxial stellar systems

    Full text link
    We review the methods used to study the orbital structure and chaotic properties of various galactic models and to construct self-consistent equilibrium solutions by Schwarzschild's orbit superposition technique. These methods are implemented in a new publicly available software tool, SMILE, which is intended to be a convenient and interactive instrument for studying a variety of 2D and 3D models, including arbitrary potentials represented by a basis-set expansion, a spherical-harmonic expansion with coefficients being smooth functions of radius (splines), or a set of fixed point masses. We also propose two new variants of Schwarzschild modelling, in which the density of each orbit is represented by the coefficients of the basis-set or spline spherical-harmonic expansion, and the orbit weights are assigned in such a way as to reproduce the coefficients of the underlying density model. We explore the accuracy of these general-purpose potential expansions and show that they may be efficiently used to approximate a wide range of analytic density models and serve as smooth representations of discrete particle sets (e.g. snapshots from an N-body simulation), for instance, for the purpose of orbit analysis of the snapshot. For the variants of Schwarzschild modelling, we use two test cases - a triaxial Dehnen model containing a central black hole, and a model re-created from an N-body snapshot obtained by a cold collapse. These tests demonstrate that all modelling approaches are capable of creating equilibrium models.Comment: MNRAS, 24 pages, 18 figures. Software is available at http://td.lpi.ru/~eugvas/smile

    A Foliated View of Transfer Learning

    Full text link
    Transfer learning considers a learning process where a new task is solved by transferring relevant knowledge from known solutions to related tasks. While this has been studied experimentally, there lacks a foundational description of the transfer learning problem that exposes what related tasks are, and how they can be exploited. In this work, we present a definition for relatedness between tasks and identify foliations as a mathematical framework to represent such relationships.Comment: 14 pages, 6 figure

    SAR-NeRF: Neural Radiance Fields for Synthetic Aperture Radar Multi-View Representation

    Full text link
    SAR images are highly sensitive to observation configurations, and they exhibit significant variations across different viewing angles, making it challenging to represent and learn their anisotropic features. As a result, deep learning methods often generalize poorly across different view angles. Inspired by the concept of neural radiance fields (NeRF), this study combines SAR imaging mechanisms with neural networks to propose a novel NeRF model for SAR image generation. Following the mapping and projection pinciples, a set of SAR images is modeled implicitly as a function of attenuation coefficients and scattering intensities in the 3D imaging space through a differentiable rendering equation. SAR-NeRF is then constructed to learn the distribution of attenuation coefficients and scattering intensities of voxels, where the vectorized form of 3D voxel SAR rendering equation and the sampling relationship between the 3D space voxels and the 2D view ray grids are analytically derived. Through quantitative experiments on various datasets, we thoroughly assess the multi-view representation and generalization capabilities of SAR-NeRF. Additionally, it is found that SAR-NeRF augumented dataset can significantly improve SAR target classification performance under few-shot learning setup, where a 10-type classification accuracy of 91.6\% can be achieved by using only 12 images per class
    • …
    corecore