Hi, everyone, I finally finished the tunning stage with the script file " moses-mert.pl" on the development corpus, but the BLEU score in run*.mert.log files is very low as listed below: ====================================== ==> run1.mert.log <== Best point: 0.132846 0.0467879 -0.644693 0.0516577 -0.0326 0.0526244 0.0383472 -0.000443235 => 0.0478053 Stopping... : [27] seconds
==> run10.mert.log <== Best point: 0.130789 0.0831878 -0.49309 0.032857 0.018328 0.0829243 0.0663681 0.0924561 => 0.0546437 Stopping... : [293] seconds ==> run11.mert.log <== Best point: 0.130738 0.0831553 -0.492897 0.0332344 0.0183208 0.0828919 0.0663422 0.09242 => 0.0546934 Stopping... : [286] seconds ==> run12.mert.log <== Best point: 0.130136 0.0827725 -0.490628 0.0330814 0.0182365 0.0825104 0.0660368 0.0965979 => 0.0547353 Stopping... : [328] seconds ==> run13.mert.log <== Best point: 0.130136 0.0827725 -0.490628 0.0330814 0.0182365 0.0825104 0.0660368 0.0965979 => 0.0547353 Stopping... : [334] seconds ==> run2.mert.log <== Best point: 0.158121 0.0928362 -0.507586 0.00834967 0.0747225 0.0630188 0.0422052 0.053161 => 0.0476075 Stopping... : [57] seconds ==> run3.mert.log <== Best point: 0.0743158 0.113324 -0.521279 0.0131133 0.0846289 0.0769264 0.0515195 0.0648931 => 0.048868 Stopping... : [87] seconds ==> run4.mert.log <== Best point: 0.115927 0.142508 -0.450438 0.0394533 0.10284 0.0641182 0.0511402 0.0335745 => 0.0489438 Stopping... : [96] seconds ==> run5.mert.log <== Best point: 0.126696 0.103291 -0.473338 0.0321721 0.0495814 0.072928 0.0589201 0.0830735 => 0.0508062 Stopping... : [123] seconds ==> run6.mert.log <== Best point: 0.127446 0.0971059 -0.453352 0.0345673 0.0387205 0.0944486 0.0538136 0.100546 => 0.052208 Stopping... : [193] seconds ==> run7.mert.log <== Best point: 0.145001 0.0566422 -0.489566 0.0468676 0.0240972 0.067952 0.0820283 0.0878453 => 0.0535599 Stopping... : [181] seconds ==> run8.mert.log <== Best point: 0.142602 0.0884247 -0.409176 0.0736586 0.0130789 0.113885 0.052161 0.107014 => 0.053494 Stopping... : [191] seconds ==> run9.mert.log <== Best point: 0.136603 0.0804459 -0.493247 0.0328674 0.0183338 0.0788154 0.067202 0.0924855 => 0.0537308 Stopping... : [272] seconds ====================================== My experiment setting: *Language model:* 1~5-gram of Chinese (Target language) *Training corpus: * Multi-UN consists about 8.7 million Chinese-English sentence pairs whose length are between 1~100 *Development corpus: * 932 Chinese sentences that have been segmented, 932 English sentences that have been tokenized *Chinese segmentation tool:* ICTLAS *English tokenizer:* tokenizer.perl I ran the training phrases with the following command: ./train-model.perl --parts 7 --mgiza --mgiza-cpus 2 --parallel --scripts-root-dir $SCRIPTS_ROOTDIR --corpus /home/loki/Downloads/moses/scripts/target/scripts-20120222-0301/training/training_corpus/clean --f eng --e chn --alignment intersect --lm 1:5:/home/loki/Downloads/irstlm-5.70.04/scripts/train.irstlm.gz I wonder the low BLEU score is due to the lack of the switch '--reordering msd-bidirectional-fe' since the moses used the default distance-based reordering model that is fairly weak or the length of sentences are too long. And I don't know if there exist a paper related to the translation of English-to-Chinese. If there do exist, then I want to refer to its BLEU score result. Any suggestion will be appreciated Best regards Moonloki
_______________________________________________ Moses-support mailing list [email protected] http://mailman.mit.edu/mailman/listinfo/moses-support
