You should consider to read the original paper as everything is well explained inside (with examples given !) :

Papineni, K., Roukos, S., Ward, T., and Zhu, W. J. (2002). "BLEU: a method for automatic evaluation of machine translation" in ACL-2002: 40th Annual meeting of the Association for Computational Linguistics pp. 311–318

http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.19.9416&rep=rep1&type=pdf

Loïc

Le 22/02/11 15:23, John Morgan a écrit :
Thanks Loic,
One more question:
To get the individual 1, 2, 3, and 4 -gram scores you divide by a
total number of corresponding ngrams.
> From reading the multi-bleu.perl code the total comes from ngrams in
the hypothesis (I think).
Do you want the total to come from the references or the hypothesis?
John





On 2/22/11, Loïc BARRAULT<[email protected]>  wrote:
Hi John,

yes this is what we want. Consider the following :
REF : the the the the
HYP : the
Choosing the max would give 4 unigram matches instead of only 1.

Cheers,

Loïc

Le 22/02/11 01:44, John Morgan a écrit :
Sorry for the empty message.
The attached file has a segment of code that I think is choosing the
minimum  ngram match count.
Is this what you want for BLEU?
Don't you want the max?
Thanks,
John




_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support
--
Loïc BARRAULT
LIUM - Equipe LST
Université du Maine




--
Loïc BARRAULT
LIUM - Equipe LST
Université du Maine

_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to