214 research outputs found

    Neural Machine Translation into Language Varieties

    Full text link
    Both research and commercial machine translation have so far neglected the importance of properly handling the spelling, lexical and grammar divergences occurring among language varieties. Notable cases are standard national varieties such as Brazilian and European Portuguese, and Canadian and European French, which popular online machine translation services are not keeping distinct. We show that an evident side effect of modeling such varieties as unique classes is the generation of inconsistent translations. In this work, we investigate the problem of training neural machine translation from English to specific pairs of language varieties, assuming both labeled and unlabeled parallel texts, and low-resource conditions. We report experiments from English to two pairs of dialects, EuropeanBrazilian Portuguese and European-Canadian French, and two pairs of standardized varieties, Croatian-Serbian and Indonesian-Malay. We show significant BLEU score improvements over baseline systems when translation into similar languages is learned as a multilingual task with shared representations.Comment: Published at EMNLP 2018: third conference on machine translation (WMT 2018

    Transfer Learning in Multilingual Neural Machine Translation with Dynamic Vocabulary

    Full text link
    We propose a method to transfer knowledge across neural machine translation (NMT) models by means of a shared dynamic vocabulary. Our approach allows to extend an initial model for a given language pair to cover new languages by adapting its vocabulary as long as new data become available (i.e., introducing new vocabulary items if they are not included in the initial model). The parameter transfer mechanism is evaluated in two scenarios: i) to adapt a trained single language NMT system to work with a new language pair and ii) to continuously add new language pairs to grow to a multilingual NMT system. In both the scenarios our goal is to improve the translation performance, while minimizing the training convergence time. Preliminary experiments spanning five languages with different training data sizes (i.e., 5k and 50k parallel sentences) show a significant performance gain ranging from +3.85 up to +13.63 BLEU in different language directions. Moreover, when compared with training an NMT model from scratch, our transfer-learning approach allows us to reach higher performance after training up to 4% of the total training steps.Comment: Published at the International Workshop on Spoken Language Translation (IWSLT), 201

    Low Resource Neural Machine Translation: A Benchmark for Five African Languages

    Full text link
    Recent advents in Neural Machine Translation (NMT) have shown improvements in low-resource language (LRL) translation tasks. In this work, we benchmark NMT between English and five African LRL pairs (Swahili, Amharic, Tigrigna, Oromo, Somali [SATOS]). We collected the available resources on the SATOS languages to evaluate the current state of NMT for LRLs. Our evaluation, comparing a baseline single language pair NMT model against semi-supervised learning, transfer learning, and multilingual modeling, shows significant performance improvements both in the En-LRL and LRL-En directions. In terms of averaged BLEU score, the multilingual approach shows the largest gains, up to +5 points, in six out of ten translation directions. To demonstrate the generalization capability of each model, we also report results on multi-domain test sets. We release the standardized experimental data and the test sets for future works addressing the challenges of NMT in under-resourced settings, in particular for the SATOS languages.Comment: Accepted for AfricaNLP workshop at ICLR 202

    Adapting Multilingual Neural Machine Translation to Unseen Languages

    Full text link
    Multilingual Neural Machine Translation (MNMT) for low-resource languages (LRL) can be enhanced by the presence of related high-resource languages (HRL), but the relatedness of HRL usually relies on predefined linguistic assumptions about language similarity. Recently, adapting MNMT to a LRL has shown to greatly improve performance. In this work, we explore the problem of adapting an MNMT model to an unseen LRL using data selection and model adaptation. In order to improve NMT for LRL, we employ perplexity to select HRL data that are most similar to the LRL on the basis of language distance. We extensively explore data selection in popular multilingual NMT settings, namely in (zero-shot) translation, and in adaptation from a multilingual pre-trained model, for both directions (LRL-en). We further show that dynamic adaptation of the model's vocabulary results in a more favourable segmentation for the LRL in comparison with direct adaptation. Experiments show reductions in training time and significant performance gains over LRL baselines, even with zero LRL data (+13.0 BLEU), up to +17.0 BLEU for pre-trained multilingual model dynamic adaptation with related data selection. Our method outperforms current approaches, such as massively multilingual models and data augmentation, on four LRL.Comment: Accepted at the 16th International Workshop on Spoken Language Translation (IWSLT), November, 201

    Multilingual Neural Machine Translation for Low-Resource Languages

    Get PDF
    In recent years, Neural Machine Translation (NMT) has been shown to be more effective than phrase-based statistical methods, thus quickly becoming the state of the art in machine translation (MT). However, NMT systems are limited in translating low-resourced languages, due to the significant amount of parallel data that is required to learn useful mappings between languages. In this work, we show how the so-called multilingual NMT can help to tackle the challenges associated with low-resourced language translation. The underlying principle of multilingual NMT is to force the creation of hidden representations of words in a shared semantic space across multiple languages, thus enabling a positive parameter transfer across languages. Along this direction, we present multilingual translation experiments with three languages (English, Italian, Romanian) covering six translation directions, utilizing both recurrent neural networks and transformer (or self-attentive) neural networks. We then focus on the zero-shot translation problem, that is how to leverage multi-lingual data in order to learn translation directions that are not covered by the available training material. To this aim, we introduce our recently proposed iterative self-training method, which incrementally improves a multilingual NMT on a zero-shot direction by just relying on monolingual data. Our results on TED talks data show that multilingual NMT outperforms conventional bilingual NMT, that the transformer NMT outperforms recurrent NMT, and that zero-shot NMT outperforms conventional pivoting methods and even matches the performance of a fully-trained bilingual system

    Improving Zero-Shot Translation of Low-Resource Languages

    Get PDF
    Recent work on multilingual neural machine translation reported competitive performance with respect to bilingual models and surprisingly good performance even on (zeroshot) translation directions not observed at training time. We investigate here a zero-shot translation in a particularly lowresource multilingual setting. We propose a simple iterative training procedure that leverages a duality of translations directly generated by the system for the zero-shot directions. The translations produced by the system (sub-optimal since they contain mixed language from the shared vocabulary), are then used together with the original parallel data to feed and iteratively re-train the multilingual network. Over time, this allows the system to learn from its own generated and increasingly better output. Our approach shows to be effective in improving the two zero-shot directions of our multilingual model. In particular, we observed gains of about 9 BLEU points over a baseline multilingual model and up to 2.08 BLEU over a pivoting mechanism using two bilingual models. Further analysis shows that there is also a slight improvement in the non-zero-shot language directions

    Lack of Macrolide Resistance in Chlamydia trachomatis after Mass Azithromycin Distributions for Trachoma

    Get PDF
    We investigated antimicrobial drug resistance in ocular Chlamydia trachomatis 18 months after 4 biannual communitywide distributions of antimicrobial drugs in a region of Ethiopia where ocular strains of C. trachomatis are highly endemic. We found no significant differences in susceptibilities to azithromycin and doxycycline in 6 posttreatment and 4 pretreatment samples

    Trends and causes of maternal mortality in Ethiopia during 1990-2013:Findings from the Global Burden of Diseases study 2013

    Get PDF
    Background: Maternal mortality is noticeably high in sub-Saharan African countries including Ethiopia. Continuous nationwide systematic evaluation and assessment of the problem helps to design appropriate policy and strategy in Ethiopia. This study aimed to investigate the trends and causes of maternal mortality in Ethiopia between 1990 and 2013. Methods: We used the Global Burden of Diseases and Risk factors (GBD) Study 2013 data that was collected from multiple sources at national and subnational levels. Spatio-temporal Gaussian Process Regression (ST-GPR) was applied to generate best estimates of maternal mortality with 95% Uncertainty Intervals (UI). Causes of death were measured using Cause of Death Ensemble modelling (CODEm). The modified UNAIDS EPP/SPECTRUM suite model was used to estimate HIV related maternal deaths. Results: In Ethiopia, a total of 16,740 (95% UI: 14,197, 19,271) maternal deaths occurred in 1990 whereas there were 15,234 (95% UI: 11,378, 19,871) maternal deaths occurred in 2013. This finding shows that Maternal Mortality Ratio (MMR) in Ethiopia was still high in the study period. There was a minimal but insignificant change of MMR over the last 23 years. The results revealed Ethiopia is below the target of Millennium Development Goals (MGDs) related to MMR. The top five causes of maternal mortality in 2013 were other direct maternal causes such as complications of anaesthesia, embolism (air, amniotic fluid, and blood clot), and the condition of peripartum cardiomyopathy (25.7%), complications of abortions (19.6%), maternal haemorrhage (12.2%), hypertensive disorders (10.3%), and maternal sepsis and other maternal infections such as influenza, malaria, tuberculosis, and hepatitis (9.6%). Most of the maternal mortality happened during the postpartum period and majority of the deaths occurred at the age group of 20-29 years. Overall trend showed that there was a decline from 708 per 100,000 live births in 1990 to 497 per 100,000 in 2013. The annual rate of change over these years was-1.6 (95% UI:-2.8 to-0.3). Conclusion: The findings of the study highlight the need for comprehensive efforts using multisectoral collaborations from stakeholders for reducing maternal mortality in Ethiopia. It is worthwhile for policies to focus on postpartum period
    • …
    corecore