4,963 research outputs found
Weakly Supervised Learning by a Confusion Matrix of Contexts
© 2019, Springer Nature Switzerland AG. Context consideration can help provide more background and related information for weakly supervised learning. The inclusion of less documented historical and environmental context in researching diabetes amongst Pima Indians uncovered reasons which were more likely to explain why some Pima Indians had much higher rates of diabetes than Caucasians, primarily due to historical, environmental and social causes rather than their specific genetic patterns or ethnicity as suggested by many medical studies. If historical and environmental factors are considered as external contexts when not included as part of a dataset for research, some forms of internal contexts may also exist inside the dataset without being declared. This paper discusses a context construction model that transforms a confusion matrix into a matrix of categorical, incremental and correlational context to emulate a kind of internal context to search for more informative patterns in order to improve weakly supervised learning from limited labeled samples for unlabeled data. When the negative and positive labeled samples and misclassification errors are compared to “happy families” and “unhappy families”, the contexts constructed by this model in the classification experiments reflected the Anna Karenina principle well - “Happy families are all alike; every unhappy family is unhappy in its own way”, an encouraging sign to further explore contexts associated with harmonizing patterns and divisive causes for knowledge discovery in a world of uncertainty
Integrating Weakly Supervised Word Sense Disambiguation into Neural Machine Translation
This paper demonstrates that word sense disambiguation (WSD) can improve
neural machine translation (NMT) by widening the source context considered when
modeling the senses of potentially ambiguous words. We first introduce three
adaptive clustering algorithms for WSD, based on k-means, Chinese restaurant
processes, and random walks, which are then applied to large word contexts
represented in a low-rank space and evaluated on SemEval shared-task data. We
then learn word vectors jointly with sense vectors defined by our best WSD
method, within a state-of-the-art NMT system. We show that the concatenation of
these vectors, and the use of a sense selection mechanism based on the weighted
average of sense vectors, outperforms several baselines including sense-aware
ones. This is demonstrated by translation on five language pairs. The
improvements are above one BLEU point over strong NMT baselines, +4% accuracy
over all ambiguous nouns and verbs, or +20% when scored manually over several
challenging words.Comment: To appear in TAC
Context-Transformer: Tackling Object Confusion for Few-Shot Detection
Few-shot object detection is a challenging but realistic scenario, where only
a few annotated training images are available for training detectors. A popular
approach to handle this problem is transfer learning, i.e., fine-tuning a
detector pretrained on a source-domain benchmark. However, such transferred
detector often fails to recognize new objects in the target domain, due to low
data diversity of training samples. To tackle this problem, we propose a novel
Context-Transformer within a concise deep transfer framework. Specifically,
Context-Transformer can effectively leverage source-domain object knowledge as
guidance, and automatically exploit contexts from only a few training images in
the target domain. Subsequently, it can adaptively integrate these relational
clues to enhance the discriminative power of detector, in order to reduce
object confusion in few-shot scenarios. Moreover, Context-Transformer is
flexibly embedded in the popular SSD-style detectors, which makes it a
plug-and-play module for end-to-end few-shot learning. Finally, we evaluate
Context-Transformer on the challenging settings of few-shot detection and
incremental few-shot detection. The experimental results show that, our
framework outperforms the recent state-of-the-art approaches.Comment: Accepted by AAAI-202
- …