6,552 research outputs found
Using Mechanical Turk to Build Machine Translation Evaluation Sets
Building machine translation (MT) test sets is a relatively expensive task. As MT becomes increasingly desired for more and more language pairs and more and more domains, it becomes necessary to build test sets for each case. In this paper, we investigate using Amazon’s Mechanical Turk (MTurk) to make MT test sets cheaply. We find that MTurk can be used to make test sets much cheaper than professionally-produced test sets. More importantly, in experiments with multiple MT systems, we find that the MTurk-produced test sets yield essentially the same conclusions regarding system performance as the professionally-produced test sets yield.This research was supported by the EuroMatrix-Plus project funded by the European Commission, by the DARPA GALE program under Contract No. HR0011-06-2-0001, and the NSF under grant IIS-0713448. Thanks to Amazon Mechanical Turk for providing a $100 credit
Translation Memory Retrieval Methods
Translation Memory (TM) systems are one of the most widely used translation
technologies. An important part of TM systems is the matching algorithm that
determines what translations get retrieved from the bank of available
translations to assist the human translator. Although detailed accounts of the
matching algorithms used in commercial systems can't be found in the
literature, it is widely believed that edit distance algorithms are used. This
paper investigates and evaluates the use of several matching algorithms,
including the edit distance algorithm that is believed to be at the heart of
most modern commercial TM systems. This paper presents results showing how well
various matching algorithms correlate with human judgments of helpfulness
(collected via crowdsourcing with Amazon's Mechanical Turk). A new algorithm
based on weighted n-gram precision that can be adjusted for translator length
preferences consistently returns translations judged to be most helpful by
translators for multiple domains and language pairs.Comment: 9 pages, 6 tables, 3 figures; appeared in Proceedings of the 14th
Conference of the European Chapter of the Association for Computational
Linguistics, April 201
Data Cleaning for XML Electronic Dictionaries via Statistical Anomaly Detection
Many important forms of data are stored digitally in XML format. Errors can
occur in the textual content of the data in the fields of the XML. Fixing these
errors manually is time-consuming and expensive, especially for large amounts
of data. There is increasing interest in the research, development, and use of
automated techniques for assisting with data cleaning. Electronic dictionaries
are an important form of data frequently stored in XML format that frequently
have errors introduced through a mixture of manual typographical entry errors
and optical character recognition errors. In this paper we describe methods for
flagging statistical anomalies as likely errors in electronic dictionaries
stored in XML format. We describe six systems based on different sources of
information. The systems detect errors using various signals in the data
including uncommon characters, text length, character-based language models,
word-based language models, tied-field length ratios, and tied-field
transliteration models. Four of the systems detect errors based on expectations
automatically inferred from content within elements of a single field type. We
call these single-field systems. Two of the systems detect errors based on
correspondence expectations automatically inferred from content within elements
of multiple related field types. We call these tied-field systems. For each
system, we provide an intuitive analysis of the type of error that it is
successful at detecting. Finally, we describe two larger-scale evaluations
using crowdsourcing with Amazon's Mechanical Turk platform and using the
annotations of a domain expert. The evaluations consistently show that the
systems are useful for improving the efficiency with which errors in XML
electronic dictionaries can be detected.Comment: 8 pages, 4 figures, 5 tables; published in Proceedings of the 2016
IEEE Tenth International Conference on Semantic Computing (ICSC), Laguna
Hills, CA, USA, pages 79-86, February 201
The Steep Road to Happily Ever After: An Analysis of Current Visual Storytelling Models
Visual storytelling is an intriguing and complex task that only recently
entered the research arena. In this work, we survey relevant work to date, and
conduct a thorough error analysis of three very recent approaches to visual
storytelling. We categorize and provide examples of common types of errors, and
identify key shortcomings in current work. Finally, we make recommendations for
addressing these limitations in the future.Comment: Accepted to the NAACL 2019 Workshop on Shortcomings in Vision and
Language (SiVL
- …