334 research outputs found
Contrastive Learning for Lifted Networks
In this work we address supervised learning of neural networks via lifted
network formulations. Lifted networks are interesting because they allow
training on massively parallel hardware and assign energy models to
discriminatively trained neural networks. We demonstrate that the training
methods for lifted networks proposed in the literature have significant
limitations and show how to use a contrastive loss to address those
limitations. We demonstrate that this contrastive training approximates
back-propagation in theory and in practice and that it is superior to the
training objective regularly used for lifted networks.Comment: 9 pages, BMVC 201
Fully Variational Noise-Contrastive Estimation
By using the underlying theory of proper scoring rules, we design a family of
noise-contrastive estimation (NCE) methods that are tractable for latent
variable models. Both terms in the underlying NCE loss, the one using data
samples and the one using noise samples, can be lower-bounded as in variational
Bayes, therefore we call this family of losses fully variational
noise-contrastive estimation. Variational autoencoders are a particular example
in this family and therefore can be also understood as separating real data
from synthetic samples using an appropriate classification loss. We further
discuss other instances in this family of fully variational NCE objectives and
indicate differences in their empirical behavior.Comment: SCIA 2023, 13 page
Driving to Opportunity: Understanding the Links among Transportation Access, Residential Outcomes, and Economic Opportunity for Housing Voucher Recipients
In the 1990s and early 2000s, the Department of Housing and Urban Development sponsored two major experiments to test whether housing choice vouchers propelled low-income households into greater economic security, the Moving to Opportunity for Fair Housing program (MTO) and the Welfare to Work Voucher program (WTW). Using data from these programs, this study examines differences in residential location and employment outcomes between voucher recipients with access to automobiles and those without. Overall, the findings underscore the positive role of automobiles in outcomes for housing voucher participants
Pareto Meets Huber: Efficiently Avoiding Poor Minima in Robust Estimation
International audienceRobust cost optimization is the task of fitting parameters to data points containing outliers. In particular, we focus on large-scale computer vision problems, such as bundle adjustment , where Non-Linear Least Square (NLLS) solvers are the current workhorse. In this context, NLLS-based state of the art algorithms have been designed either to quickly improve the target objective and find a local minimum close to the initial value of the parameters, or to have a strong ability to avoid poor local minima. In this paper, we propose a novel algorithm relying on multi-objective optimization which allows to match those two properties. We experimentally demonstrate that our algorithm has an ability to avoid poor local minima that is on par with the best performing algorithms with a faster decrease of the target objective
Lifted Regression/Reconstruction Networks
In this work we propose lifted regression/reconstruction networks (LRRNs),
which combine lifted neural networks with a guaranteed Lipschitz continuity
property for the output layer. Lifted neural networks explicitly optimize an
energy model to infer the unit activations and therefore---in contrast to
standard feed-forward neural networks---allow bidirectional feedback between
layers. So far lifted neural networks have been modelled around standard
feed-forward architectures. We propose to take further advantage of the
feedback property by letting the layers simultaneously perform regression and
reconstruction. The resulting lifted network architecture allows to control the
desired amount of Lipschitz continuity, which is an important feature to obtain
adversarially robust regression and classification methods. We analyse and
numerically demonstrate applications for unsupervised and supervised learning.Comment: 12 pages, 8 figure
Seeing Behind Things: Extending Semantic Segmentation to Occluded Regions
Semantic segmentation and instance level segmentation made substantial
progress in recent years due to the emergence of deep neural networks (DNNs). A
number of deep architectures with Convolution Neural Networks (CNNs) were
proposed that surpass the traditional machine learning approaches for
segmentation by a large margin. These architectures predict the directly
observable semantic category of each pixel by usually optimizing a cross
entropy loss. In this work we push the limit of semantic segmentation towards
predicting semantic labels of directly visible as well as occluded objects or
objects parts, where the network's input is a single depth image. We group the
semantic categories into one background and multiple foreground object groups,
and we propose a modification of the standard cross-entropy loss to cope with
the settings. In our experiments we demonstrate that a CNN trained by
minimizing the proposed loss is able to predict semantic categories for visible
and occluded object parts without requiring to increase the network size
(compared to a standard segmentation task). The results are validated on a
newly generated dataset (augmented from SUNCG) dataset
- …