Hello All,

I am a student. I am new to MT. I have configured moses baseline system.
For English-Punjabi language pair. It is working fine. Now, I wish to train
factored SMT. For this I have trained 2000 sentences in
surface-word|lemma|POS-tag format for both sides and will train another
2000.

My queries are
1. would 4000 sentences would be enough.
2. What memory overheads would be used.
3. Would training a pos based lm requires a file with seperate only pos tag.

Please answer my queries.

Kabir.
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to