176,695 research outputs found
Minimizing Supervision in Multi-label Categorization
Multiple categories of objects are present in most images. Treating this as a
multi-class classification is not justified. We treat this as a multi-label
classification problem. In this paper, we further aim to minimize the
supervision required for providing supervision in multi-label classification.
Specifically, we investigate an effective class of approaches that associate a
weak localization with each category either in terms of the bounding box or
segmentation mask. Doing so improves the accuracy of multi-label
categorization. The approach we adopt is one of active learning, i.e.,
incrementally selecting a set of samples that need supervision based on the
current model, obtaining supervision for these samples, retraining the model
with the additional set of supervised samples and proceeding again to select
the next set of samples. A crucial concern is the choice of the set of samples.
In doing so, we provide a novel insight, and no specific measure succeeds in
obtaining a consistently improved selection criterion. We, therefore, provide a
selection criterion that consistently improves the overall baseline criterion
by choosing the top k set of samples for a varied set of criteria. Using this
criterion, we are able to show that we can retain more than 98% of the fully
supervised performance with just 20% of samples (and more than 96% using 10%)
of the dataset on PASCAL VOC 2007 and 2012. Also, our proposed approach
consistently outperforms all other baseline metrics for all benchmark datasets
and model combinations.Comment: Accepted in CVPR-W 202
Adversarial Data Programming: Using GANs to Relax the Bottleneck of Curated Labeled Data
Paucity of large curated hand-labeled training data for every
domain-of-interest forms a major bottleneck in the deployment of machine
learning models in computer vision and other fields. Recent work (Data
Programming) has shown how distant supervision signals in the form of labeling
functions can be used to obtain labels for given data in near-constant time. In
this work, we present Adversarial Data Programming (ADP), which presents an
adversarial methodology to generate data as well as a curated aggregated label
has given a set of weak labeling functions. We validated our method on the
MNIST, Fashion MNIST, CIFAR 10 and SVHN datasets, and it outperformed many
state-of-the-art models. We conducted extensive experiments to study its
usefulness, as well as showed how the proposed ADP framework can be used for
transfer learning as well as multi-task learning, where data from two domains
are generated simultaneously using the framework along with the label
information. Our future work will involve understanding the theoretical
implications of this new framework from a game-theoretic perspective, as well
as explore the performance of the method on more complex datasets.Comment: CVPR 2018 main conference pape
CALDA: Improving Multi-Source Time Series Domain Adaptation with Contrastive Adversarial Learning
Unsupervised domain adaptation (UDA) provides a strategy for improving
machine learning performance in data-rich (target) domains where ground truth
labels are inaccessible but can be found in related (source) domains. In cases
where meta-domain information such as label distributions is available, weak
supervision can further boost performance. We propose a novel framework, CALDA,
to tackle these two problems. CALDA synergistically combines the principles of
contrastive learning and adversarial learning to robustly support multi-source
UDA (MS-UDA) for time series data. Similar to prior methods, CALDA utilizes
adversarial learning to align source and target feature representations. Unlike
prior approaches, CALDA additionally leverages cross-source label information
across domains. CALDA pulls examples with the same label close to each other,
while pushing apart examples with different labels, reshaping the space through
contrastive learning. Unlike prior contrastive adaptation methods, CALDA
requires neither data augmentation nor pseudo labeling, which may be more
challenging for time series. We empirically validate our proposed approach.
Based on results from human activity recognition, electromyography, and
synthetic datasets, we find utilizing cross-source information improves
performance over prior time series and contrastive methods. Weak supervision
further improves performance, even in the presence of noise, allowing CALDA to
offer generalizable strategies for MS-UDA. Code is available at:
https://github.com/floft/caldaComment: Under review at IEEE Transactions on Pattern Analysis and Machine
Intelligenc
Multi-Instance Multilabel Learning with Weak-Label for Predicting Protein Function in Electricigens
Nature often brings several domains together to form multidomain and multifunctional proteins with a vast number of possibilities. In our previous study, we disclosed that the protein function prediction problem is naturally and inherently Multi-Instance Multilabel (MIML) learning tasks. Automated protein function prediction is typically implemented under the assumption that the functions of labeled proteins are complete; that is, there are no missing labels. In contrast, in practice just a subset of the functions of a protein are known, and whether this protein has other functions is unknown. It is evident that protein function prediction tasks suffer from weak-label problem; thus protein function prediction with incomplete annotation matches well with the MIML with weak-label learning framework. In this paper, we have applied the state-of-the-art MIML with weak-label learning algorithm MIMLwel for predicting protein functions in two typical real-world electricigens organisms which have been widely used in microbial fuel cells (MFCs) researches. Our experimental results validate the effectiveness of MIMLwel algorithm in predicting protein functions with incomplete annotation
- …