126,132 research outputs found
Classification-reconstruction learning for open-set recognition
Open-set classification is a problem of handling `unknown' classes that are
not contained in the training dataset, whereas traditional classifiers assume
that only known classes appear in the test environment. Existing open-set
classifiers rely on deep networks trained in a supervised manner on known
classes in the training set; this causes specialization of learned
representations to known classes and makes it hard to distinguish unknowns from
knowns. In contrast, we train networks for joint classification and
reconstruction of input data. This enhances the learned representation so as to
preserve information useful for separating unknowns from knowns, as well as to
discriminate classes of knowns. Our novel Classification-Reconstruction
learning for Open-Set Recognition (CROSR) utilizes latent representations for
reconstruction and enables robust unknown detection without harming the
known-class classification accuracy. Extensive experiments reveal that the
proposed method outperforms existing deep open-set classifiers in multiple
standard datasets and is robust to diverse outliers. The code is available in
https://nae-lab.org/~rei/research/crosr/.Comment: 11 pages, 7 figure
C2AE: Class Conditioned Auto-Encoder for Open-set Recognition
Models trained for classification often assume that all testing classes are
known while training. As a result, when presented with an unknown class during
testing, such closed-set assumption forces the model to classify it as one of
the known classes. However, in a real world scenario, classification models are
likely to encounter such examples. Hence, identifying those examples as unknown
becomes critical to model performance. A potential solution to overcome this
problem lies in a class of learning problems known as open-set recognition. It
refers to the problem of identifying the unknown classes during testing, while
maintaining performance on the known classes. In this paper, we propose an
open-set recognition algorithm using class conditioned auto-encoders with novel
training and testing methodology. In contrast to previous methods, training
procedure is divided in two sub-tasks, 1. closed-set classification and, 2.
open-set identification (i.e. identifying a class as known or unknown). Encoder
learns the first task following the closed-set classification training
pipeline, whereas decoder learns the second task by reconstructing conditioned
on class identity. Furthermore, we model reconstruction errors using the
Extreme Value Theory of statistical modeling to find the threshold for
identifying known/unknown class samples. Experiments performed on multiple
image classification datasets show proposed method performs significantly
better than state of the art.Comment: CVPR2019 (Oral
Positive-unlabeled learning for open set domain adaptation
Open Set Domain Adaptation (OSDA) focuses on bridging the domain gap between a labeled source domain and an unlabeled target domain, while also rejecting target classes that are not present in the source as unknown. The challenges of this task are closely related to those of Positive-Unlabeled (PU) learning where it is essential to discriminate between positive (known) and negative (unknown) class samples in the unlabeled target data. With this newly discovered connection, we leverage the theoretical framework of PU learning for OSDA and, at the same time, we extend PU learning to tackle uneven data distributions. Our method combines domain adversarial learning with a new non-negative risk estimator for PU learning based on self-supervised sample reconstruction. With experiments on digit recognition and object classification, we validate our risk estimator and demonstrate that our approach allows reducing the domain gap without suffering from negative transfer
Generative Model with Coordinate Metric Learning for Object Recognition Based on 3D Models
Given large amount of real photos for training, Convolutional neural network
shows excellent performance on object recognition tasks. However, the process
of collecting data is so tedious and the background are also limited which
makes it hard to establish a perfect database. In this paper, our generative
model trained with synthetic images rendered from 3D models reduces the
workload of data collection and limitation of conditions. Our structure is
composed of two sub-networks: semantic foreground object reconstruction network
based on Bayesian inference and classification network based on multi-triplet
cost function for avoiding over-fitting problem on monotone surface and fully
utilizing pose information by establishing sphere-like distribution of
descriptors in each category which is helpful for recognition on regular photos
according to poses, lighting condition, background and category information of
rendered images. Firstly, our conjugate structure called generative model with
metric learning utilizing additional foreground object channels generated from
Bayesian rendering as the joint of two sub-networks. Multi-triplet cost
function based on poses for object recognition are used for metric learning
which makes it possible training a category classifier purely based on
synthetic data. Secondly, we design a coordinate training strategy with the
help of adaptive noises acting as corruption on input images to help both
sub-networks benefit from each other and avoid inharmonious parameter tuning
due to different convergence speed of two sub-networks. Our structure achieves
the state of the art accuracy of over 50\% on ShapeNet database with data
migration obstacle from synthetic images to real photos. This pipeline makes it
applicable to do recognition on real images only based on 3D models.Comment: 14 page
- …