How many bits are needed to store probabilities for phrasebased translation

Abstract

State of the art in statistical machine trans-lation is currently represented by phrase-based models, which typically incorpo-rate a large number of probabilities of phrase-pairs and word n-grams. In this work, we investigate data compression methods for efciently encoding n-gram and phrase-pair probabilities, that are usu-ally encoded in 32-bit oating point num-bers. We measured the impact of com-pression on translation quality through a phrase-based decoder trained on two dis-tinct tasks: the translation of European Parliament speeches from Spanish to En-glish, and the translation of news agencies from Chinese to English. We show that with a very simple quantization scheme all probabilities can be encoded in just 4 bits with a relative loss in BLEU score on the two tasks by 1.0 % and 1.6%, respectively.

Similar works

Full text

thumbnail-image

CiteSeerX

redirect
Last time updated on 28/10/2017

This paper was published in CiteSeerX.

Having an issue?

Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.