69 research outputs found

    Joint Energy-based Modelling for Remote Sensing Image Processing

    Get PDF
    The peta-scale, continuously increasing amount of publicly available remote sensing information forms an unprecedented archive of Earth observation data. Although advances in deep learning provide tools to exploit big amounts of digital information, most supervised methods rely on accurately annotated sets to train models. Access to large amounts of high-quality annotations proves costly due to the human labor involved. Such limitations have been studied in semi-supervised learning where unlabeled samples aid the generalization of models trained with limited amounts of labeled data. The Joint Energy-based Model (JEM) is a recent, physics-inspired approach simultaneously optimizing a supervised task along with a generative process to train a sampler approximating a data distribution. Although a promising formulation of such models, current JEM implementations are predominantly applied to classification tasks. Their potential improving semantic segmentation tasks remains locked. Our work investigates JEM training behavior from a conceptual perspective, studying mechanisms of loss function divergences that numerically destabilizes the model optimization. We explore three regularization terms imposed on energy values and optimization gradients to alleviate the training complexity. Our experiments indicate that the proposed regularization mitigates loss function divergences for remote sensing imagery classification. Regularization on energy values of real samples performed the best. Additionally, we present an extended definition of JEM for image segmentation, sJEM. In our experiments, the generation branch did not perform as expected. sJEM was unable to generate realistic remote-sensing-like samples. Correspondingly performance is biased for the sJEM segmentation branch. Initial model optimization runs demand additional research to stabilize the methodology given spatial auto-correlations in remote sensing multi-spectral imagery. Our insights pave the way for the design of follow-up research to advance sJEM for Earth observation

    InfoNCE is a variational autoencoder

    Full text link
    We show that a popular self-supervised learning method, InfoNCE, is a special case of a new family of unsupervised learning methods, the self-supervised variational autoencoder (SSVAE). SSVAEs circumvent the usual VAE requirement to reconstruct the data by using a carefully chosen implicit decoder. The InfoNCE objective was motivated as a simplified parametric mutual information estimator. Under one choice of prior, the SSVAE objective (i.e. the ELBO) is exactly equal to the mutual information (up to constants). Under an alternative choice of prior, the SSVAE objective is exactly equal to the simplified parametric mutual information estimator used in InfoNCE (up to constants). Importantly, the use of simplified parametric mutual information estimators is believed to be critical to obtain good high-level representations, and the SSVAE framework naturally provides a principled justification for using prior information to choose these estimators

    Lifted Regression/Reconstruction Networks

    Get PDF
    In this work we propose lifted regression/reconstruction networks(LRRNs), which combine lifted neural networks with a guaranteed Lipschitz continuity property for the output layer. Lifted neural networks explicitly optimize an energy model to infer the unit activations and therefore—in contrast to standard feed-forward neural networks—allow bidirectional feedback between layers. So far lifted neural networks have been modelled around standard feed-forward architectures. We propose to take further advantage of the feedback property by letting the layers simultaneously perform regression and reconstruction. The resulting lifted network architecture allows to control the desired amount of Lipschitz continuity, which is an important feature to obtain adversarially robust regression and classification methods. We analyse and numerically demonstrate applications for unsupervised and supervised learnin
    • …
    corecore