373 research outputs found
Capturing Label Characteristics in VAEs
We present a principled approach to incorporating labels in VAEs that
captures the rich characteristic information associated with those labels.
While prior work has typically conflated these by learning latent variables
that directly correspond to label values, we argue this is contrary to the
intended effect of supervision in VAEs-capturing rich label characteristics
with the latents. For example, we may want to capture the characteristics of a
face that make it look young, rather than just the age of the person. To this
end, we develop the CCVAE, a novel VAE model and concomitant variational
objective which captures label characteristics explicitly in the latent space,
eschewing direct correspondences between label values and latents. Through
judicious structuring of mappings between such characteristic latents and
labels, we show that the CCVAE can effectively learn meaningful representations
of the characteristics of interest across a variety of supervision schemes. In
particular, we show that the CCVAE allows for more effective and more general
interventions to be performed, such as smooth traversals within the
characteristics for a given label, diverse conditional generation, and
transferring characteristics across datapoints.Comment: Accepted to ICLR 202
Resampling Generative Models: An Empirical Study
Despite Generative AI’s rapid growth (e.g., ChatGPT, GPT-4, Dalle-2, etc.), generated data from these models may have an inherent bias. This bias can propagate to downstream tasks e.g., classification, and data augmentation that utilize data from generative AI models. This thesis empirically evaluates model bias in different deep generative models like Variational Autoencoder and PixelCNN++. Further, we resample generated data using importance sampling to reduce bias in the generated images based on a recently proposed method for bias-reduction using probabilistic classifiers. The approach is developed in the context of image generation and we demonstrate that importance sampling can produce better quality samples with lower bias. Next, we improve downstream classification by developing a semi-supervised learning pipeline where we use importance-sampled data as unlabeled examples within a classifier. Specifically, we use a loss function called as the semantic-loss function that was proposed to add constraints on unlabeled data to improve the performance of classification using limited labeled examples. Through the use of importance-sampled images, we essentially add constraints on data instances that are more informative for the classifier, thus resulting in the classifier learning a better decision boundary using fewer labeled examples
InfoNCE is a variational autoencoder
We show that a popular self-supervised learning method, InfoNCE, is a special
case of a new family of unsupervised learning methods, the self-supervised
variational autoencoder (SSVAE). SSVAEs circumvent the usual VAE requirement to
reconstruct the data by using a carefully chosen implicit decoder. The InfoNCE
objective was motivated as a simplified parametric mutual information
estimator. Under one choice of prior, the SSVAE objective (i.e. the ELBO) is
exactly equal to the mutual information (up to constants). Under an alternative
choice of prior, the SSVAE objective is exactly equal to the simplified
parametric mutual information estimator used in InfoNCE (up to constants).
Importantly, the use of simplified parametric mutual information estimators is
believed to be critical to obtain good high-level representations, and the
SSVAE framework naturally provides a principled justification for using prior
information to choose these estimators
- …