You should consider to read the original paper as everything is well
explained inside (with examples given !) :
Papineni, K., Roukos, S., Ward, T., and Zhu, W. J. (2002). "BLEU: a
method for automatic evaluation of machine translation" in ACL-2002:
40th Annual meeting of the Association for Computational Linguistics pp.
311–318
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.19.9416&rep=rep1&type=pdf
Loïc
Le 22/02/11 15:23, John Morgan a écrit :
Thanks Loic,
One more question:
To get the individual 1, 2, 3, and 4 -gram scores you divide by a
total number of corresponding ngrams.
> From reading the multi-bleu.perl code the total comes from ngrams in
the hypothesis (I think).
Do you want the total to come from the references or the hypothesis?
John
On 2/22/11, Loïc BARRAULT<[email protected]> wrote:
Hi John,
yes this is what we want. Consider the following :
REF : the the the the
HYP : the
Choosing the max would give 4 unigram matches instead of only 1.
Cheers,
Loïc
Le 22/02/11 01:44, John Morgan a écrit :
Sorry for the empty message.
The attached file has a segment of code that I think is choosing the
minimum ngram match count.
Is this what you want for BLEU?
Don't you want the max?
Thanks,
John
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support
--
Loïc BARRAULT
LIUM - Equipe LST
Université du Maine
--
Loïc BARRAULT
LIUM - Equipe LST
Université du Maine
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support