167 research outputs found
Trivial Transfer Learning for Low-Resource Neural Machine Translation
Transfer learning has been proven as an effective technique for neural
machine translation under low-resource conditions. Existing methods require a
common target language, language relatedness, or specific training tricks and
regimes. We present a simple transfer learning method, where we first train a
"parent" model for a high-resource language pair and then continue the training
on a lowresource pair only by replacing the training corpus. This "child" model
performs significantly better than the baseline trained for lowresource pair
only. We are the first to show this for targeting different languages, and we
observe the improvements even for unrelated languages with different alphabets.Comment: Accepted to WMT18 reseach paper, Proceedings of the 3rd Conference on
Machine Translation 201
Transfer Learning for Low-Resource Part-of-Speech Tagging
Neural network approaches to Part-of-Speech tagging, like other supervised neural network tasks, benefit from larger quantities of labeled data. However, in the case of low-resource languages, additional methods are necessary to improve the performances of POS taggers. In this paper, we explore transfer learning approaches to improve POS tagging in Afrikaans using a neural network. We investigate the effect of transferring network weights that were originally trained for POS tagging in Dutch. We also test the use of pretrained word embeddings in our POS tagger, both independently and in conjunction with the transferred weights from a Dutch POS tagger. We find a marginal increase in performance due to transfer learning with the Dutch POS tagger, and a significant increase due to the use of either unaligned or aligned pretrained embeddings. Notably, there is little difference in performance when using either unaligned or aligned embeddings, even when utilizing cross-lingual transfer learning
Data Augmentation for Low-Resource Neural Machine Translation
The quality of a Neural Machine Translation system depends substantially on
the availability of sizable parallel corpora. For low-resource language pairs
this is not the case, resulting in poor translation quality. Inspired by work
in computer vision, we propose a novel data augmentation approach that targets
low-frequency words by generating new sentence pairs containing rare words in
new, synthetically created contexts. Experimental results on simulated
low-resource settings show that our method improves translation quality by up
to 2.9 BLEU points over the baseline and up to 3.2 BLEU over back-translation.Comment: 5 pages, 2 figures, Accepted at ACL 201
Low Resource Neural Machine Translation: A Benchmark for Five African Languages
Recent advents in Neural Machine Translation (NMT) have shown improvements in
low-resource language (LRL) translation tasks. In this work, we benchmark NMT
between English and five African LRL pairs (Swahili, Amharic, Tigrigna, Oromo,
Somali [SATOS]). We collected the available resources on the SATOS languages to
evaluate the current state of NMT for LRLs. Our evaluation, comparing a
baseline single language pair NMT model against semi-supervised learning,
transfer learning, and multilingual modeling, shows significant performance
improvements both in the En-LRL and LRL-En directions. In terms of averaged
BLEU score, the multilingual approach shows the largest gains, up to +5 points,
in six out of ten translation directions. To demonstrate the generalization
capability of each model, we also report results on multi-domain test sets. We
release the standardized experimental data and the test sets for future works
addressing the challenges of NMT in under-resourced settings, in particular for
the SATOS languages.Comment: Accepted for AfricaNLP workshop at ICLR 202
- …