5 research outputs found
It's LeVAsa not LevioSA! Latent Encodings for Valence-Arousal Structure Alignment
In recent years, great strides have been made in the field of affective
computing. Several models have been developed to represent and quantify
emotions. Two popular ones include (i) categorical models which represent
emotions as discrete labels, and (ii) dimensional models which represent
emotions in a Valence-Arousal (VA) circumplex domain. However, there is no
standard for annotation mapping between the two labelling methods. We build a
novel algorithm for mapping categorical and dimensional model labels using
annotation transfer across affective facial image datasets. Further, we utilize
the transferred annotations to learn rich and interpretable data
representations using a variational autoencoder (VAE). We present "LeVAsa", a
VAE model that learns implicit structure by aligning the latent space with the
VA space. We evaluate the efficacy of LeVAsa by comparing performance with the
Vanilla VAE using quantitative and qualitative analysis on two benchmark
affective image datasets. Our results reveal that LeVAsa achieves high
latent-circumplex alignment which leads to improved downstream categorical
emotion prediction. The work also demonstrates the trade-off between degree of
alignment and quality of reconstructions.Comment: 5 pages, 4 figures and 3 table
Imitative Planning using Conditional Normalizing Flow
We explore the application of normalizing flows for improving the performance
of trajectory planning for autonomous vehicles (AVs). Normalizing flows provide
an invertible mapping from a known prior distribution to a potentially complex,
multi-modal target distribution and allow for fast sampling with exact PDF
inference. By modeling a trajectory planner's cost manifold as an energy
function we learn a scene conditioned mapping from the prior to a Boltzmann
distribution over the AV control space. This mapping allows for control samples
and their associated energy to be generated jointly and in parallel. We propose
using neural autoregressive flow (NAF) as part of an end-to-end deep learned
system that allows for utilizing sensors, map, and route information to
condition the flow mapping. Finally, we demonstrate the effectiveness of our
approach on real world datasets over IL and hand constructed trajectory
sampling techniques.Comment: Submittted to 4th Conference on Robot Learning (CoRL 2020), Cambridge
MA, US
Multi-Rate VAE: Train Once, Get the Full Rate-Distortion Curve
Variational autoencoders (VAEs) are powerful tools for learning latent
representations of data used in a wide range of applications. In practice, VAEs
usually require multiple training rounds to choose the amount of information
the latent variable should retain. This trade-off between the reconstruction
error (distortion) and the KL divergence (rate) is typically parameterized by a
hyperparameter . In this paper, we introduce Multi-Rate VAE (MR-VAE), a
computationally efficient framework for learning optimal parameters
corresponding to various in a single training run. The key idea is to
explicitly formulate a response function that maps to the optimal
parameters using hypernetworks. MR-VAEs construct a compact response
hypernetwork where the pre-activations are conditionally gated based on
. We justify the proposed architecture by analyzing linear VAEs and
showing that it can represent response functions exactly for linear VAEs. With
the learned hypernetwork, MR-VAEs can construct the rate-distortion curve
without additional training and can be deployed with significantly less
hyperparameter tuning. Empirically, our approach is competitive and often
exceeds the performance of multiple -VAEs training with minimal
computation and memory overheads.Comment: 22 pages, 9 figure