13 research outputs found

    The AMU-UEDIN Submission to the WMT16 News Translation Task: Attention-based NMT Models as Feature Functions in Phrase-based SMT

    Get PDF
    This paper describes the AMU-UEDIN submissions to the WMT 2016 shared task on news translation. We explore methods of decode-time integration of attention-based neural translation models with phrase-based statistical machine translation. Efficient batch-algorithms for GPU-querying are proposed and implemented. For English-Russian, our system stays behind the state-of-the-art pure neural models in terms of BLEU. Among restricted systems, manual evaluation places it in the first cluster tied with the pure neural model. For the Russian-English task, our submission achieves the top BLEU result, outperforming the best pure neural system by 1.1 BLEU points and our own phrase-based baseline by 1.6 BLEU. After manual evaluation, this system is the best restricted system in its own cluster. In follow-up experiments we improve results by additional 0.8 BLEU

    Is Neural Machine Translation Ready for Deployment? A Case Study on 30 Translation Directions

    Get PDF
    In this paper we provide the largest published comparison of translation quality for phrase-based SMT and neural machine translation across 30 translation directions. For ten directions we also include hierarchical phrase-based MT. Experiments are performed for the recently published United Nations Parallel Corpus v1.0 and its large six-way sentence-aligned subcorpus. In the second part of the paper we investigate aspects of translation speed, introducing AmuNMT, our efficient neural machine translation decoder. We demonstrate that current neural machine translation could already be used for in-production systems when comparing words-per-second ratios.Comment: Accepted for presentation at IWSLT 2016, Seattl

    Fast Neural Machine Translation Implementation

    Get PDF
    This paper describes the submissions to the efficiency track for GPUs at the Workshop for Neural Machine Translation and Generation by members of the University of Edinburgh, Adam Mickiewicz University, Tilde and University of Alicante. We focus on efficient implementation of the recurrent deep-learning model as implemented in Amun, the fast inference engine for neural machine translation. We improve the performance with an efficient mini-batching algorithm, and by fusing the softmax operation with the k-best extraction algorithm. Submissions using Amun were first, second and third fastest in the GPU efficiency track

    Predicting Target Language CCG Supertags Improves Neural Machine Translation

    Get PDF
    Neural machine translation (NMT) models are able to partially learn syntactic information from sequential lexical information. Still, some complex syntactic phenomena such as prepositional phrase attachment are poorly modeled. This work aims to answer two questions: 1) Does explicitly modeling target language syntax help NMT? 2) Is tight integration of words and syntax better than multitask training? We introduce syntactic information in the form of CCG supertags in the decoder, by interleaving the target supertags with the word sequence. Our results on WMT data show that explicitly modeling target-syntax improves machine translation quality for German->English, a high-resource pair, and for Romanian->English, a low-resource pair and also several syntactic phenomena including prepositional phrase attachment. Furthermore, a tight coupling of words and syntax improves translation quality more than multitask training. By combining target-syntax with adding source-side dependency labels in the embedding layer, we obtain a total improvement of 0.9 BLEU for German->English and 1.2 BLEU for Romanian->English.Comment: Accepted at the Second Conference on Machine Translation (WMT17). This version includes more results regarding target syntax for Romanian->English and reports fewer results regarding source synta

    Marian: Fast Neural Machine Translation in C++

    Get PDF
    We present Marian, an efficient and self-contained Neural Machine Translation framework with an integrated automatic differentiation engine based on dynamic computation graphs. Marian is written entirely in C++. We describe the design of the encoder-decoder framework and demonstrate that a research-friendly toolkit can achieve high training and translation speed.Comment: Demonstration pape

    Neurological symptoms in hospitalised patients with COVID-19 and their association with in-hospital mortality

    Get PDF
    Objectives. To evaluate the spectrum of neurological symptoms in patients with COVID-19 during the first 14 days of hospitalisation and its association with in-hospital mortality. Material and methods. We included 200 patients with RT-PCR-confirmed COVID-19 admitted to University Hospital in Krakow, Poland. In 164 patients, a detailed questionnaire concerning neurological symptoms and signs was performed prospectively within 14 days of hospitalisation. In the remaining 36 patients, such questionnaires were completed retrospectively based on daily observations in the Department of Neurology. Results. During hospitalisation, 169 patients (84.5%) experienced neurological symptoms; the most common were: fatigue (62.5%), decreased mood (45.5%), myalgia (43.5%), and muscle weakness (42.5%). Patients who died during hospitalisation compared to the remainder were older (79 [70.5–88.5] vs. 63.5 [51–77] years, p = 0.001), and more often had decreased level of consciousness (50.0% vs. 9.3%, p < 0.001), delirium (33.3% vs. 4.4%, p < 0.001), arterial hypotension (50.0% vs. 19.6%, p = 0.005) or stroke during (18.8% vs. 3.3%, p = 0.026) or before hospitalisation (50.0% vs. 7.1, p < 0.001), whereas those who survived more often suffered from headache (42.1% vs. 0%, p = 0.012) or decreased mood (51.7% vs. 0%, p = 0.003). Conclusions. Most hospitalised patients with COVID-19 experience neurological symptoms. Decreased level of consciousness, delirium, arterial hypotension, and stroke during or before hospitalisation increase the risk of in-hospital mortality

    The SUMMA Platform Prototype

    Get PDF
    We present the first prototype of the SUMMA Platform: an integrated platform for multilingual media monitoring. The platform contains a rich suite of low-level and high-level natural language processing technologies: automatic speech recognition of broadcast media, machine translation, automated tagging and classification of named entities, semantic parsing to detect relationships between entities, and automatic construction / augmentation of factual knowledge bases. Implemented on the Docker platform, it can easily be deployed, customised, and scaled to large volumes of incoming media streams

    From dataset recycling to multi-property extraction and beyond

    No full text
    This paper investigates various Transformerarchitectures on the WikiReading Informa-tion Extraction and Machine Reading Com-prehension dataset. The proposed dual-sourcemodel outperforms the current state-of-the-art by a large margin.Next, we intro-duce WikiReading Recycled—a newly devel-oped public dataset, and the task of multiple-property extraction. It uses the same data asWikiReading but does not inherit its predeces-sor’s identified disadvantages. In addition, weprovide a human-annotated test set with diag-nostic subsets for a detailed analysis of modelperformance

    Going full-TILT boogie on document understanding with text-image-layout transformer

    No full text
    We address the challenging problem of Natural Language Comprehension beyond plain-text documents by introducing the TILT neural network architecture which simultaneously learns layout information, visual features, and textual semantics. Contrary to previous approaches, we rely on a decoder capable of unifying a variety of problems involving natural language. The layout is represented as an attention bias and complemented with contextualized visual information, while the core of our model is a pretrained encoder-decoder Transformer. Our novel approach achieves state-of-the-art results in extracting information from documents and answering questions which demand layout understanding (DocVQA, CORD, SROIE). At the same time, we simplify the process by employing an end-to-end model.Comment: Accepted at ICDAR 202
    corecore