Hi everyone!

We all know that MERT gets slower in the later iterations. This is not 
surprising as the n-best lists of all previous iterations are merged. I believe 
this is quite important for translation performance.

Still, it seems important to me to get the merged lists as small as possible. A 
quick inspection of mert/extractor indicates that duplicates are _not_ removed. 
Can anyone confirm this? And is this really not done anywhere else, e.g. in 
mert/mert ?

Removing duplicates in the extractor should be easy to implement and I don't 
think it will take more running time than one gains from smaller list.

Best,
 Thomas (currently University of Pisa)
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to