66 research outputs found
Improved Robust Algorithms for Learning with Discriminative Feature Feedback
Discriminative Feature Feedback is a setting proposed by Dastupta et al.
(2018), which provides a protocol for interactive learning based on feature
explanations that are provided by a human teacher. The features distinguish
between the labels of pairs of possibly similar instances. That work has shown
that learning in this model can have considerable statistical and computational
advantages over learning in standard label-based interactive learning models.
In this work, we provide new robust interactive learning algorithms for the
Discriminative Feature Feedback model, with mistake bounds that are
significantly lower than those of previous robust algorithms for this setting.
In the adversarial setting, we reduce the dependence on the number of protocol
exceptions from quadratic to linear. In addition, we provide an algorithm for a
slightly more restricted model, which obtains an even smaller mistake bound for
large models with many exceptions.
In the stochastic setting, we provide the first algorithm that converges to
the exception rate with a polynomial sample complexity. Our algorithm and
analysis for the stochastic setting involve a new construction that we call
Feature Influence, which may be of wider applicability.Comment: AISTATS 202
Active Nearest-Neighbor Learning in Metric Spaces
We propose a pool-based non-parametric active learning algorithm for general
metric spaces, called MArgin Regularized Metric Active Nearest Neighbor
(MARMANN), which outputs a nearest-neighbor classifier. We give prediction
error guarantees that depend on the noisy-margin properties of the input
sample, and are competitive with those obtained by previously proposed passive
learners. We prove that the label complexity of MARMANN is significantly lower
than that of any passive learner with similar error guarantees. MARMANN is
based on a generalized sample compression scheme, and a new label-efficient
active model-selection procedure
Fast Single-Class Classification and the Principle of Logit Separation
We consider neural network training, in applications in which there are many
possible classes, but at test-time, the task is a binary classification task of
determining whether the given example belongs to a specific class, where the
class of interest can be different each time the classifier is applied. For
instance, this is the case for real-time image search. We define the Single
Logit Classification (SLC) task: training the network so that at test-time, it
would be possible to accurately identify whether the example belongs to a given
class in a computationally efficient manner, based only on the output logit for
this class. We propose a natural principle, the Principle of Logit Separation,
as a guideline for choosing and designing losses suitable for the SLC. We show
that the cross-entropy loss function is not aligned with the Principle of Logit
Separation. In contrast, there are known loss functions, as well as novel batch
loss functions that we propose, which are aligned with this principle. In
total, we study seven loss functions. Our experiments show that indeed in
almost all cases, losses that are aligned with the Principle of Logit
Separation obtain at least 20% relative accuracy improvement in the SLC task
compared to losses that are not aligned with it, and sometimes considerably
more. Furthermore, we show that fast SLC does not cause any drop in binary
classification accuracy, compared to standard classification in which all
logits are computed, and yields a speedup which grows with the number of
classes. For instance, we demonstrate a 10x speedup when the number of classes
is 400,000. Tensorflow code for optimizing the new batch losses is publicly
available at https://github.com/cruvadom/Logit Separation.Comment: Published as a conference paper in ICDM 201
- …