18 research outputs found
Improving Neural Relation Extraction with Positive and Unlabeled Learning
We present a novel approach to improve the performance of distant supervision
relation extraction with Positive and Unlabeled (PU) Learning. This approach
first applies reinforcement learning to decide whether a sentence is positive
to a given relation, and then positive and unlabeled bags are constructed. In
contrast to most previous studies, which mainly use selected positive instances
only, we make full use of unlabeled instances and propose two new
representations for positive and unlabeled bags. These two representations are
then combined in an appropriate way to make bag-level prediction. Experimental
results on a widely used real-world dataset demonstrate that this new approach
indeed achieves significant and consistent improvements as compared to several
competitive baselines.Comment: 8 pages, AAAI-202
Relaxed Softmax for learning from Positive and Unlabeled data
In recent years, the softmax model and its fast approximations have become
the de-facto loss functions for deep neural networks when dealing with
multi-class prediction. This loss has been extended to language modeling and
recommendation, two fields that fall into the framework of learning from
Positive and Unlabeled data. In this paper, we stress the different drawbacks
of the current family of softmax losses and sampling schemes when applied in a
Positive and Unlabeled learning setup. We propose both a Relaxed Softmax loss
(RS) and a new negative sampling scheme based on Boltzmann formulation. We show
that the new training objective is better suited for the tasks of density
estimation, item similarity and next-event prediction by driving uplifts in
performance on textual and recommendation datasets against classical softmax.Comment: 9 pages, 5 figures, 2 tables, published at RecSys 201
A Symmetric Loss Perspective of Reliable Machine Learning
When minimizing the empirical risk in binary classification, it is a common
practice to replace the zero-one loss with a surrogate loss to make the
learning objective feasible to optimize. Examples of well-known surrogate
losses for binary classification include the logistic loss, hinge loss, and
sigmoid loss. It is known that the choice of a surrogate loss can highly
influence the performance of the trained classifier and therefore it should be
carefully chosen. Recently, surrogate losses that satisfy a certain symmetric
condition (aka., symmetric losses) have demonstrated their usefulness in
learning from corrupted labels. In this article, we provide an overview of
symmetric losses and their applications. First, we review how a symmetric loss
can yield robust classification from corrupted labels in balanced error rate
(BER) minimization and area under the receiver operating characteristic curve
(AUC) maximization. Then, we demonstrate how the robust AUC maximization method
can benefit natural language processing in the problem where we want to learn
only from relevant keywords and unlabeled documents. Finally, we conclude this
article by discussing future directions, including potential applications of
symmetric losses for reliable machine learning and the design of non-symmetric
losses that can benefit from the symmetric condition.Comment: Preprint of an Invited Review Articl
Iterative multi-path tracking for video and volume segmentation with sparse point supervision
Recent machine learning strategies for segmentation tasks have shown great
ability when trained on large pixel-wise annotated image datasets. It remains a
major challenge however to aggregate such datasets, as the time and monetary
cost associated with collecting extensive annotations is extremely high. This
is particularly the case for generating precise pixel-wise annotations in video
and volumetric image data. To this end, this work presents a novel framework to
produce pixel-wise segmentations using minimal supervision. Our method relies
on 2D point supervision, whereby a single 2D location within an object of
interest is provided on each image of the data. Our method then estimates the
object appearance in a semi-supervised fashion by learning
object-image-specific features and by using these in a semi-supervised learning
framework. Our object model is then used in a graph-based optimization problem
that takes into account all provided locations and the image data in order to
infer the complete pixel-wise segmentation. In practice, we solve this
optimally as a tracking problem using a K-shortest path approach. Both the
object model and segmentation are then refined iteratively to further improve
the final segmentation. We show that by collecting 2D locations using a gaze
tracker, our approach can provide state-of-the-art segmentations on a range of
objects and image modalities (video and 3D volumes), and that these can then be
used to train supervised machine learning classifiers