1 research outputs found

    Human Post-editing in Hybrid Machine Translation Systems: Automatic and Manual Analysis and Evaluation

    Get PDF
    This study assesses, automatically and manually, the performance of two hybrid machine translation (HMT) systems, via a text corpus of questions in the Spanish and English languages. The results show that human evaluation metrics are more reliable when evaluating HMT performance. Further, there is evidence that MT can streamline the translation process for specific types of texts, such as questions; however, it does not yet rival the quality of human translations, to which post-editing is key in this process
    corecore