6 research outputs found
Análisis de los puntos claves en la documentación para una buena traducción médica (ES - EN)
La presente investigación desarrolló como objetivo general analizar los puntos
claves para lograr una buena documentación en la traducción médica. Con
respecto a la metodología, el estudio es de enfoque cualitativo, de tipo aplicado,
con diseño de teoría fundamentada y nivel descriptivo. La técnica que se empleó
fue el análisis de documentos, y el instrumento fue la ficha de análisis. La muestra
estuvo compuesta por 5 autores que propusieron 5 diferentes técnicas de
documentación y por 2 fuentes de donde se obtuvieron las características del
lenguaje médico y por lo tanto las técnicas de recolección de datos enfocados en
la traducción médica. Los resultados revelaron que, dentro de los muchos autores
estudiados, existen 5 técnicas especialmente para la traducción que se pueden
adaptar a la rama de la traducción médica y también lenguaje especializado que
puede llegar a ser muy útil para la creación de nuevas propuestas de recolección
de información. Finalmente se concluye que existen 3 puntos clave que se deberían
tomar en cuenta, y con esa base, adaptar las técnicas a nuestra práctica
profesional, estos 3 puntos clave tienen una estructura piramidal muy marcada,
estos son: la documentación general, la documentación especializada – médica y
la documentación en base al criterio profesional de los aspectos externos al
encargo de traducción
Low-resource speech translation
We explore the task of speech-to-text translation (ST), where speech in one language
(source) is converted to text in a different one (target). Traditional ST systems go
through an intermediate step where the source language speech is first converted to
source language text using an automatic speech recognition (ASR) system, which
is then converted to target language text using a machine translation (MT) system.
However, this pipeline based approach is impractical for unwritten languages spoken by
millions of people around the world, leaving them without access to free and automated
translation services such as Google Translate. The lack of such translation services can
have important real-world consequences. For example, in the aftermath of a disaster
scenario, easily available translation services can help better co-ordinate relief efforts.
How can we expand the coverage of automated ST systems to include scenarios which
lack source language text? In this thesis we investigate one possible solution: we
build ST systems to directly translate source language speech into target language text,
thereby forgoing the dependency on source language text. To build such a system, we
use only speech data paired with text translations as training data. We also specifically
focus on low-resource settings, where we expect at most tens of hours of training data
to be available for unwritten or endangered languages.
Our work can be broadly divided into three parts. First we explore how we can leverage
prior work to build ST systems. We find that neural sequence-to-sequence models are
an effective and convenient method for ST, but produce poor quality translations when
trained in low-resource settings.
In the second part of this thesis, we explore methods to improve the translation performance
of our neural ST systems which do not require labeling additional speech
data in the low-resource language, a potentially tedious and expensive process. Instead
we exploit labeled speech data for high-resource languages which is widely available
and relatively easier to obtain. We show that pretraining a neural model with ASR data
from a high-resource language, different from both the source and target ST languages,
improves ST performance.
In the final part of our thesis, we study whether ST systems can be used to build
applications which have traditionally relied on the availability of ASR systems, such
as information retrieval, clustering audio documents, or question/answering. We build
proof-of-concept systems for two downstream applications: topic prediction for speech
and cross-lingual keyword spotting. Our results indicate that low-resource ST systems
can still outperform simple baselines for these tasks, leaving the door open for further
exploratory work.
This thesis provides, for the first time, an in-depth study of neural models for the
task of direct ST across a range of training data settings on a realistic multi-speaker
speech corpus. Our contributions include a set of open-source tools to encourage further
research