15 research outputs found
Multi-space Variational Encoder-Decoders for Semi-supervised Labeled Sequence Transduction
Labeled sequence transduction is a task of transforming one sequence into
another sequence that satisfies desiderata specified by a set of labels. In
this paper we propose multi-space variational encoder-decoders, a new model for
labeled sequence transduction with semi-supervised learning. The generative
model can use neural networks to handle both discrete and continuous latent
variables to exploit various features of data. Experiments show that our model
provides not only a powerful supervised framework but also can effectively take
advantage of the unlabeled data. On the SIGMORPHON morphological inflection
benchmark, our model outperforms single-model state-of-art results by a large
margin for the majority of languages.Comment: Accepted by ACL 201
Improved English to Russian Translation by Neural Suffix Prediction
Neural machine translation (NMT) suffers a performance deficiency when a
limited vocabulary fails to cover the source or target side adequately, which
happens frequently when dealing with morphologically rich languages. To address
this problem, previous work focused on adjusting translation granularity or
expanding the vocabulary size. However, morphological information is relatively
under-considered in NMT architectures, which may further improve translation
quality. We propose a novel method, which can not only reduce data sparsity but
also model morphology through a simple but effective mechanism. By predicting
the stem and suffix separately during decoding, our system achieves an
improvement of up to 1.98 BLEU compared with previous work on English to
Russian translation. Our method is orthogonal to different NMT architectures
and stably gains improvements on various domains.Comment: 8 pages, 3 figures, 5 table
Translating into Morphologically Rich Languages with Synthetic Phrases
Translation into morphologically rich languages is an important but recalcitrant problem in MT. We present a simple and effective approach that deals with the problem in two phases. First, a discriminative model is learned to predict inflections of target words from rich source-side annotations. Then, this model is used to create additional sentencespecific word- and phrase-level translations that are added to a standard translation model as “synthetic” phrases. Our approach relies on morphological analysis of the target language, but we show that an unsupervised Bayesian model of morphology can successfully be used in place of a supervised analyzer. We report significant improvements in translation quality when translating from English to Russian, Hebrew and Swahili.</p