6 research outputs found
Multi-Level Generative Models for Partial Label Learning with Non-random Label Noise
Partial label (PL) learning tackles the problem where each training instance
is associated with a set of candidate labels that include both the true label
and irrelevant noise labels. In this paper, we propose a novel multi-level
generative model for partial label learning (MGPLL), which tackles the problem
by learning both a label level adversarial generator and a feature level
adversarial generator under a bi-directional mapping framework between the
label vectors and the data samples. Specifically, MGPLL uses a conditional
noise label generation network to model the non-random noise labels and perform
label denoising, and uses a multi-class predictor to map the training instances
to the denoised label vectors, while a conditional data feature generator is
used to form an inverse mapping from the denoised label vectors to data
samples. Both the noise label generator and the data feature generator are
learned in an adversarial manner to match the observed candidate labels and
data features respectively. Extensive experiments are conducted on synthesized
and real-world partial label datasets. The proposed approach demonstrates the
state-of-the-art performance for partial label learning
Combating noisy labels by agreement: A joint training method with co-regularization
Deep Learning with noisy labels is a practically challenging problem in
weakly supervised learning. The state-of-the-art approaches "Decoupling" and
"Co-teaching+" claim that the "disagreement" strategy is crucial for
alleviating the problem of learning with noisy labels. In this paper, we start
from a different perspective and propose a robust learning paradigm called
JoCoR, which aims to reduce the diversity of two networks during training.
Specifically, we first use two networks to make predictions on the same
mini-batch data and calculate a joint loss with Co-Regularization for each
training example. Then we select small-loss examples to update the parameters
of both two networks simultaneously. Trained by the joint loss, these two
networks would be more and more similar due to the effect of Co-Regularization.
Extensive experimental results on corrupted data from benchmark datasets
including MNIST, CIFAR-10, CIFAR-100 and Clothing1M demonstrate that JoCoR is
superior to many state-of-the-art approaches for learning with noisy labels.Comment: Accepted by CVPR 2020; Code is available at:
https://github.com/hongxin001/JoCoR. arXiv admin note: text overlap with
arXiv:1901.04215 by other author
HERA: Partial Label Learning by Combining Heterogeneous Loss with Sparse and Low-Rank Regularization
Partial Label Learning (PLL) aims to learn from the data where each training
instance is associated with a set of candidate labels, among which only one is
correct. Most existing methods deal with such problem by either treating each
candidate label equally or identifying the ground-truth label iteratively. In
this paper, we propose a novel PLL approach called HERA, which simultaneously
incorporates the HeterogEneous Loss and the SpaRse and Low-rAnk procedure to
estimate the labeling confidence for each instance while training the model.
Specifically, the heterogeneous loss integrates the strengths of both the
pairwise ranking loss and the pointwise reconstruction loss to provide
informative label ranking and reconstruction information for label
identification, while the embedded sparse and low-rank scheme constrains the
sparsity of ground-truth label matrix and the low rank of noise label matrix to
explore the global label relevance among the whole training data for improving
the learning model. Extensive experiments on both artificial and real-world
data sets demonstrate that our method can achieve superior or comparable
performance against the state-of-the-art methods
Incorporating Multiple Cluster Centers for Multi-Label Learning
Multi-label learning deals with the problem that each instance is associated
with multiple labels simultaneously. Most of the existing approaches aim to
improve the performance of multi-label learning by exploiting label
correlations. Although the data augmentation technique is widely used in many
machine learning tasks, it is still unclear whether data augmentation is
helpful to multi-label learning. In this paper, (to the best of our knowledge)
we provide the first attempt to leverage the data augmentation technique to
improve the performance of multi-label learning. Specifically, we first propose
a novel data augmentation approach that performs clustering on the real
examples and treats the cluster centers as virtual examples, and these virtual
examples naturally embody the local label correlations and label importances.
Then, motivated by the cluster assumption that examples in the same cluster
should have the same label, we propose a novel regularization term to bridge
the gap between the real examples and virtual examples, which can promote the
local smoothness of the learning function. Extensive experimental results on a
number of real-world multi-label data sets clearly demonstrate that our
proposed approach outperforms the state-of-the-art counterparts.Comment: 19 pages with 4 figures and 4 table
GM-PLL: Graph Matching based Partial Label Learning
Partial Label Learning (PLL) aims to learn from the data where each training
example is associated with a set of candidate labels, among which only one is
correct. The key to deal with such problem is to disambiguate the candidate
label sets and obtain the correct assignments between instances and their
candidate labels. In this paper, we interpret such assignments as
instance-to-label matchings, and reformulate the task of PLL as a matching
selection problem. To model such problem, we propose a novel Graph Matching
based Partial Label Learning (GM-PLL) framework, where Graph Matching (GM)
scheme is incorporated owing to its excellent capability of exploiting the
instance and label relationship. Meanwhile, since conventional one-to-one GM
algorithm does not satisfy the constraint of PLL problem that multiple
instances may correspond to the same label, we extend a traditional one-to-one
probabilistic matching algorithm to the many-to-one constraint, and make the
proposed framework accommodate to the PLL problem. Moreover, we also propose a
relaxed matching prediction model, which can improve the prediction accuracy
via GM strategy. Extensive experiments on both artificial and real-world data
sets demonstrate that the proposed method can achieve superior or comparable
performance against the state-of-the-art methods
Provably Consistent Partial-Label Learning
Partial-label learning (PLL) is a multi-class classification problem, where
each training example is associated with a set of candidate labels. Even though
many practical PLL methods have been proposed in the last two decades, there
lacks a theoretical understanding of the consistency of those methods-none of
the PLL methods hitherto possesses a generation process of candidate label
sets, and then it is still unclear why such a method works on a specific
dataset and when it may fail given a different dataset. In this paper, we
propose the first generation model of candidate label sets, and develop two
novel PLL methods that are guaranteed to be provably consistent, i.e., one is
risk-consistent and the other is classifier-consistent. Our methods are
advantageous, since they are compatible with any deep network or stochastic
optimizer. Furthermore, thanks to the generation model, we would be able to
answer the two questions above by testing if the generation model matches given
candidate label sets. Experiments on benchmark and real-world datasets validate
the effectiveness of the proposed generation model and two PLL methods