69 research outputs found
Joint Energy-based Modelling for Remote Sensing Image Processing
The peta-scale, continuously increasing amount of publicly available remote sensing information forms an unprecedented archive of Earth observation data. Although advances in deep learning provide tools to exploit big amounts of digital information, most supervised methods rely on accurately annotated sets to train models. Access to large amounts of high-quality annotations proves costly due to the human labor involved. Such limitations have been studied in semi-supervised learning where unlabeled samples aid the generalization of models trained with limited amounts of labeled data. The Joint Energy-based Model (JEM) is a recent, physics-inspired approach simultaneously optimizing a supervised task along with a generative process to train a sampler approximating a data distribution. Although a promising formulation of such models, current JEM implementations are predominantly applied to classification tasks. Their potential improving semantic segmentation tasks remains locked.
Our work investigates JEM training behavior from a conceptual perspective, studying mechanisms of loss function divergences that numerically destabilizes the model optimization. We explore three regularization terms imposed on energy values and optimization gradients to alleviate the training complexity. Our experiments indicate that the proposed regularization mitigates loss function divergences for remote sensing imagery classification. Regularization on energy values of real samples performed the best.
Additionally, we present an extended definition of JEM for image segmentation, sJEM. In our experiments, the generation branch did not perform as expected. sJEM was unable to generate realistic remote-sensing-like samples. Correspondingly performance is biased for the sJEM segmentation branch. Initial model optimization runs demand additional research to stabilize the methodology given spatial auto-correlations in remote sensing multi-spectral imagery. Our insights pave the way for the design of follow-up research to advance sJEM for Earth observation
InfoNCE is a variational autoencoder
We show that a popular self-supervised learning method, InfoNCE, is a special
case of a new family of unsupervised learning methods, the self-supervised
variational autoencoder (SSVAE). SSVAEs circumvent the usual VAE requirement to
reconstruct the data by using a carefully chosen implicit decoder. The InfoNCE
objective was motivated as a simplified parametric mutual information
estimator. Under one choice of prior, the SSVAE objective (i.e. the ELBO) is
exactly equal to the mutual information (up to constants). Under an alternative
choice of prior, the SSVAE objective is exactly equal to the simplified
parametric mutual information estimator used in InfoNCE (up to constants).
Importantly, the use of simplified parametric mutual information estimators is
believed to be critical to obtain good high-level representations, and the
SSVAE framework naturally provides a principled justification for using prior
information to choose these estimators
Lifted Regression/Reconstruction Networks
In this work we propose lifted regression/reconstruction networks(LRRNs), which combine lifted neural networks with a guaranteed Lipschitz continuity property for the output layer. Lifted neural networks explicitly optimize an energy model to infer the unit activations and therefore—in contrast to standard feed-forward neural networks—allow bidirectional feedback between layers. So far lifted neural networks have been modelled around standard feed-forward architectures. We propose to take further advantage of the feedback property by letting the layers simultaneously perform regression and reconstruction. The resulting lifted network architecture allows to control the desired amount of Lipschitz continuity, which is an important feature to obtain adversarially robust regression and classification methods. We analyse and numerically demonstrate applications for unsupervised and supervised learnin
- …