607 research outputs found
Striking the Right Balance with Uncertainty
Learning unbiased models on imbalanced datasets is a significant challenge.
Rare classes tend to get a concentrated representation in the classification
space which hampers the generalization of learned boundaries to new test
examples. In this paper, we demonstrate that the Bayesian uncertainty estimates
directly correlate with the rarity of classes and the difficulty level of
individual samples. Subsequently, we present a novel framework for uncertainty
based class imbalance learning that follows two key insights: First,
classification boundaries should be extended further away from a more uncertain
(rare) class to avoid overfitting and enhance its generalization. Second, each
sample should be modeled as a multi-variate Gaussian distribution with a mean
vector and a covariance matrix defined by the sample's uncertainty. The learned
boundaries should respect not only the individual samples but also their
distribution in the feature space. Our proposed approach efficiently utilizes
sample and class uncertainty information to learn robust features and more
generalizable classifiers. We systematically study the class imbalance problem
and derive a novel loss formulation for max-margin learning based on Bayesian
uncertainty measure. The proposed method shows significant performance
improvements on six benchmark datasets for face verification, attribute
prediction, digit/object classification and skin lesion detection.Comment: CVPR 201
Wrapped Cauchy Distributed Angular Softmax for Long-Tailed Visual Recognition
Addressing imbalanced or long-tailed data is a major challenge in visual
recognition tasks due to disparities between training and testing distributions
and issues with data noise. We propose the Wrapped Cauchy Distributed Angular
Softmax (WCDAS), a novel softmax function that incorporates data-wise
Gaussian-based kernels into the angular correlation between feature
representations and classifier weights, effectively mitigating noise and sparse
sampling concerns. The class-wise distribution of angular representation
becomes a sum of these kernels. Our theoretical analysis reveals that the
wrapped Cauchy distribution excels the Gaussian distribution in approximating
mixed distributions. Additionally, WCDAS uses trainable concentration
parameters to dynamically adjust the compactness and margin of each class.
Empirical results confirm label-aware behavior in these parameters and
demonstrate WCDAS's superiority over other state-of-the-art softmax-based
methods in handling long-tailed visual recognition across multiple benchmark
datasets. The code is public available.Comment: accepted by ICML 202
Recurrent Pixel Embedding for Instance Grouping
We introduce a differentiable, end-to-end trainable framework for solving
pixel-level grouping problems such as instance segmentation consisting of two
novel components. First, we regress pixels into a hyper-spherical embedding
space so that pixels from the same group have high cosine similarity while
those from different groups have similarity below a specified margin. We
analyze the choice of embedding dimension and margin, relating them to
theoretical results on the problem of distributing points uniformly on the
sphere. Second, to group instances, we utilize a variant of mean-shift
clustering, implemented as a recurrent neural network parameterized by kernel
bandwidth. This recurrent grouping module is differentiable, enjoys convergent
dynamics and probabilistic interpretability. Backpropagating the group-weighted
loss through this module allows learning to focus on only correcting embedding
errors that won't be resolved during subsequent clustering. Our framework,
while conceptually simple and theoretically abundant, is also practically
effective and computationally efficient. We demonstrate substantial
improvements over state-of-the-art instance segmentation for object proposal
generation, as well as demonstrating the benefits of grouping loss for
classification tasks such as boundary detection and semantic segmentation
Zero-Shot Learning -- A Comprehensive Evaluation of the Good, the Bad and the Ugly
Due to the importance of zero-shot learning, i.e. classifying images where
there is a lack of labeled training data, the number of proposed approaches has
recently increased steadily. We argue that it is time to take a step back and
to analyze the status quo of the area. The purpose of this paper is three-fold.
First, given the fact that there is no agreed upon zero-shot learning
benchmark, we first define a new benchmark by unifying both the evaluation
protocols and data splits of publicly available datasets used for this task.
This is an important contribution as published results are often not comparable
and sometimes even flawed due to, e.g. pre-training on zero-shot test classes.
Moreover, we propose a new zero-shot learning dataset, the Animals with
Attributes 2 (AWA2) dataset which we make publicly available both in terms of
image features and the images themselves. Second, we compare and analyze a
significant number of the state-of-the-art methods in depth, both in the
classic zero-shot setting but also in the more realistic generalized zero-shot
setting. Finally, we discuss in detail the limitations of the current status of
the area which can be taken as a basis for advancing it.Comment: Accepted by TPAMI in July, 2018. We introduce Proposed Split Version
2.0 (Please download it from our project webpage). arXiv admin note:
substantial text overlap with arXiv:1703.0439
- …