6 research outputs found

    Análisis de los puntos claves en la documentación para una buena traducción médica (ES - EN)

    Get PDF
    La presente investigación desarrolló como objetivo general analizar los puntos claves para lograr una buena documentación en la traducción médica. Con respecto a la metodología, el estudio es de enfoque cualitativo, de tipo aplicado, con diseño de teoría fundamentada y nivel descriptivo. La técnica que se empleó fue el análisis de documentos, y el instrumento fue la ficha de análisis. La muestra estuvo compuesta por 5 autores que propusieron 5 diferentes técnicas de documentación y por 2 fuentes de donde se obtuvieron las características del lenguaje médico y por lo tanto las técnicas de recolección de datos enfocados en la traducción médica. Los resultados revelaron que, dentro de los muchos autores estudiados, existen 5 técnicas especialmente para la traducción que se pueden adaptar a la rama de la traducción médica y también lenguaje especializado que puede llegar a ser muy útil para la creación de nuevas propuestas de recolección de información. Finalmente se concluye que existen 3 puntos clave que se deberían tomar en cuenta, y con esa base, adaptar las técnicas a nuestra práctica profesional, estos 3 puntos clave tienen una estructura piramidal muy marcada, estos son: la documentación general, la documentación especializada – médica y la documentación en base al criterio profesional de los aspectos externos al encargo de traducción

    Low-resource speech translation

    Get PDF
    We explore the task of speech-to-text translation (ST), where speech in one language (source) is converted to text in a different one (target). Traditional ST systems go through an intermediate step where the source language speech is first converted to source language text using an automatic speech recognition (ASR) system, which is then converted to target language text using a machine translation (MT) system. However, this pipeline based approach is impractical for unwritten languages spoken by millions of people around the world, leaving them without access to free and automated translation services such as Google Translate. The lack of such translation services can have important real-world consequences. For example, in the aftermath of a disaster scenario, easily available translation services can help better co-ordinate relief efforts. How can we expand the coverage of automated ST systems to include scenarios which lack source language text? In this thesis we investigate one possible solution: we build ST systems to directly translate source language speech into target language text, thereby forgoing the dependency on source language text. To build such a system, we use only speech data paired with text translations as training data. We also specifically focus on low-resource settings, where we expect at most tens of hours of training data to be available for unwritten or endangered languages. Our work can be broadly divided into three parts. First we explore how we can leverage prior work to build ST systems. We find that neural sequence-to-sequence models are an effective and convenient method for ST, but produce poor quality translations when trained in low-resource settings. In the second part of this thesis, we explore methods to improve the translation performance of our neural ST systems which do not require labeling additional speech data in the low-resource language, a potentially tedious and expensive process. Instead we exploit labeled speech data for high-resource languages which is widely available and relatively easier to obtain. We show that pretraining a neural model with ASR data from a high-resource language, different from both the source and target ST languages, improves ST performance. In the final part of our thesis, we study whether ST systems can be used to build applications which have traditionally relied on the availability of ASR systems, such as information retrieval, clustering audio documents, or question/answering. We build proof-of-concept systems for two downstream applications: topic prediction for speech and cross-lingual keyword spotting. Our results indicate that low-resource ST systems can still outperform simple baselines for these tasks, leaving the door open for further exploratory work. This thesis provides, for the first time, an in-depth study of neural models for the task of direct ST across a range of training data settings on a realistic multi-speaker speech corpus. Our contributions include a set of open-source tools to encourage further research
    corecore