4 research outputs found
Correlated Logistic Model With Elastic Net Regularization for Multilabel Image Classification
In this paper, we present correlated logistic (CorrLog) model for multilabel
image classification. CorrLog extends conventional logistic regression model
into multilabel cases, via explicitly modeling the pairwise correlation between
labels. In addition, we propose to learn the model parameters of CorrLog with
elastic net regularization, which helps exploit the sparsity in feature
selection and label correlations and thus further boost the performance of
multilabel classification. CorrLog can be efficiently learned, though
approximately, by regularized maximum pseudo likelihood estimation, and it
enjoys a satisfying generalization bound that is independent of the number of
labels. CorrLog performs competitively for multilabel image classification on
benchmark data sets MULAN scene, MIT outdoor scene, PASCAL VOC 2007, and PASCAL
VOC 2012, compared with the state-of-the-art multilabel classification
algorithms
On the benefits of output sparsity for multi-label classification
The multi-label classification framework, where each observation can be
associated with a set of labels, has generated a tremendous amount of attention
over recent years. The modern multi-label problems are typically large-scale in
terms of number of observations, features and labels, and the amount of labels
can even be comparable with the amount of observations. In this context,
different remedies have been proposed to overcome the curse of dimensionality.
In this work, we aim at exploiting the output sparsity by introducing a new
loss, called the sparse weighted Hamming loss. This proposed loss can be seen
as a weighted version of classical ones, where active and inactive labels are
weighted separately. Leveraging the influence of sparsity in the loss function,
we provide improved generalization bounds for the empirical risk minimizer, a
suitable property for large-scale problems. For this new loss, we derive rates
of convergence linear in the underlying output-sparsity rather than linear in
the number of labels. In practice, minimizing the associated risk can be
performed efficiently by using convex surrogates and modern convex optimization
algorithms. We provide experiments on various real-world datasets demonstrating
the pertinence of our approach when compared to non-weighted techniques
A Hidden Variables Approach to Multilabel Logistic Regression
Multilabel classification is an important problem in a wide range of domains
such as text categorization and music annotation. In this paper, we present a
probabilistic model, Multilabel Logistic Regression with Hidden variables
(MLRH), which extends the standard logistic regression by introducing hidden
variables. Hidden variables make it possible to go beyond the conventional
multiclass logistic regression by relaxing the one-hot-encoding constraint. We
define a new joint distribution of labels and hidden variables which enables us
to obtain one classifier for multilabel classification. Our experimental
studies on a set of benchmark datasets demonstrate that the probabilistic model
can achieve competitive performance compared with other multilabel learning
algorithms.Comment: 16 page
Reconstruction Regularized Deep Metric Learning for Multi-label Image Classification
In this paper, we present a novel deep metric learning method to tackle the
multi-label image classification problem. In order to better learn the
correlations among images features, as well as labels, we attempt to explore a
latent space, where images and labels are embedded via two unique deep neural
networks, respectively. To capture the relationships between image features and
labels, we aim to learn a \emph{two-way} deep distance metric over the
embedding space from two different views, i.e., the distance between one image
and its labels is not only smaller than those distances between the image and
its labels' nearest neighbors, but also smaller than the distances between the
labels and other images corresponding to the labels' nearest neighbors.
Moreover, a reconstruction module for recovering correct labels is incorporated
into the whole framework as a regularization term, such that the label
embedding space is more representative. Our model can be trained in an
end-to-end manner. Experimental results on publicly available image datasets
corroborate the efficacy of our method compared with the state-of-the-arts.Comment: Accepted by IEEE TNNL