Hi,
I have trained my dataset using the following script;
~/build/mosesdecoder/scripts/training/train-model.perl -root-dir
train \
-corpus ~/factored_translation/dataset/pos/train
\
--f tr --e en -alignment tr-en.align -reordering
msd-bidirectional-fe \
-lm 0:3:/export/students/sait/factored_translation/lm/surface.lm:8
\
-lm 2:3:/export/students/sait/factored_translation/lm/pos.lm:8
\
--translation-factors 0-0,2
\
--external-bin-dir ~/build/mosesdecoder/bin
My input token factors are;
- surface form,
- lemma,
- part of speech, and
- part of speech with additional morphological information
My output token factors are;
- surface form,
- lemma, and
- part of speech
I have created two different language models. surface.lm is the model of
words in the target corpus, and pos.lm is the model of pos tags in the
target corpus.
When I translate, I obtain such sentences:
*-->They had to is cash from some bootleg DVDs can see are hiding her *
*-->cat ! did you get any Yoon is the one I BRANDON BRANDON I think I am
ever let me fires his songs *
*-->constant by idolise They have it invcible a sister . that it contender
They had to have ever happened . her into the in my heart it Mum *
*-->is tired without leaving francis recently aII-out he ca n't and thrusts
out the words ever put either of famous case long , oh do got some *
*-->BROWN : You dead father carry 's manager dead parents . BRENNER
Youbetterwatchout 's manager They had to have just moved my he 's going to
lose high volume ever had an STD ? mulling*
Is 0.5 BLEU score normal, what do you think? I mean, of course, when we
look at sentences, it is quite normal. But is it normal with the parameters
I have given for training?
Kind Regards,
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support