9,045 research outputs found
Low-shot learning with large-scale diffusion
This paper considers the problem of inferring image labels from images when
only a few annotated examples are available at training time. This setup is
often referred to as low-shot learning, where a standard approach is to
re-train the last few layers of a convolutional neural network learned on
separate classes for which training examples are abundant. We consider a
semi-supervised setting based on a large collection of images to support label
propagation. This is possible by leveraging the recent advances on large-scale
similarity graph construction.
We show that despite its conceptual simplicity, scaling label propagation up
to hundred millions of images leads to state of the art accuracy in the
low-shot learning regime
Semi-Supervised Learning, Causality and the Conditional Cluster Assumption
While the success of semi-supervised learning (SSL) is still not fully
understood, Sch\"olkopf et al. (2012) have established a link to the principle
of independent causal mechanisms. They conclude that SSL should be impossible
when predicting a target variable from its causes, but possible when predicting
it from its effects. Since both these cases are somewhat restrictive, we extend
their work by considering classification using cause and effect features at the
same time, such as predicting disease from both risk factors and symptoms.
While standard SSL exploits information contained in the marginal distribution
of all inputs (to improve the estimate of the conditional distribution of the
target given inputs), we argue that in our more general setting we should use
information in the conditional distribution of effect features given causal
features. We explore how this insight generalises the previous understanding,
and how it relates to and can be exploited algorithmically for SSL.Comment: 36th Conference on Uncertainty in Artificial Intelligence (2020)
(Previously presented at the NeurIPS 2019 workshop "Do the right thing":
machine learning and causal inference for improved decision making,
Vancouver, Canada.
Unsupervised Domain Adaptation using Graph Transduction Games
Unsupervised domain adaptation (UDA) amounts to assigning class labels to the
unlabeled instances of a dataset from a target domain, using labeled instances
of a dataset from a related source domain. In this paper, we propose to cast
this problem in a game-theoretic setting as a non-cooperative game and
introduce a fully automatized iterative algorithm for UDA based on graph
transduction games (GTG). The main advantages of this approach are its
principled foundation, guaranteed termination of the iterative algorithms to a
Nash equilibrium (which corresponds to a consistent labeling condition) and
soft labels quantifying the uncertainty of the label assignment process. We
also investigate the beneficial effect of using pseudo-labels from linear
classifiers to initialize the iterative process. The performance of the
resulting methods is assessed on publicly available object recognition
benchmark datasets involving both shallow and deep features. Results of
experiments demonstrate the suitability of the proposed game-theoretic approach
for solving UDA tasks.Comment: Oral IJCNN 201
- …