On 07/05/2015 20:43, Alla Rozovskaya wrote:
Hi Hieu,

Please see below.

(1) Thank you for your help. I am using a standard LM training for now, without the N-gram corpus. Could you please let me know how I can train a translation model without any reordering model (I did not find anything in the manual about it). what's the exact command you're using to create the translation model and reordering model?

$mosesDir/scripts/training/train-model.perl -root-dir $outputDir/mosesFiles/ -corpus $corpusDir/$trainFile.clean -f $source -e $target -alignment grow-diag-final-and -reordering msd-bidirectional-fe -lm 0:$order:$lm1 -external-bin-dir /scratch/rozovska/mosesdecoder/tools/ -cores 2 --max-phrase-length 7

I don't know how to prevent the creation of the re-ordering model. However, if you don't want to use it in decoding, then just delete the LexicalReordering and Distortion model entries in the moses.ini file BEFORE you tune.

I tried removing Distortion and LexicalReordering from the moses.ini file but I still found these cases in the translated file.

(2) Also, can I use multiple language models when training a translation model, or is it only possible to specify multiple LMs in the moses.ini file? the LM isn't used when creating the translation model. They're only used in decoding

What is the -lm parameter in train-model.perl used for?
the -lm parameter is just used to add the LM model entry to the moses.ini file.

In fact, if you don't like the moses.ini that is created automatically for you, create your own that has just the models you need and use that for tuning and decoding

Thanks a lot for your help,

Alla

On Wed, May 6, 2015 at 12:12 AM, Hieu Hoang <[email protected] <mailto:[email protected]>> wrote:



    Hieu Hoang
    Researcher
    New York University, Abu Dhabi
    http://www.hoang.co.uk/hieu

    On 6 May 2015 at 01:36, Alla Rozovskaya <[email protected]
    <mailto:[email protected]>> wrote:

        Hi Hieu,

        Thank you for your help. I am using a standard LM training for
        now, without the N-gram corpus. Could you please let me know
        how I can train a translation model without any reordering
        model (I did not find anything in the manual about it).

    what's the exact command you're using to create the translation
    model and reordering model?

        Also, can I use multiple language models when training a
        translation model, or is it only possible to specify multiple
        LMs in the moses.ini file?

    the LM isn't used when creating the translation model. They're
    only used in decoding


        Thanks again,

        Alla

        On Tue, Apr 28, 2015 at 10:06 AM, Hieu Hoang
        <[email protected] <mailto:[email protected]>> wrote:

            I spoke to Ken about using KenLM to train a standard
            backoff LM with the n-gram corpus. It's not supported yet
            or recommended.

            I'm not sure whether the moses' SRILM wrapper will support
            the count-based LM. And how much memory it will consume.
            Try it and please let us know.

            People have also been using the Common Crawl corpus to
            build huge backoff LM. They're very difficult to use as it
            consumes a lot of memory


            On 25/04/2015 20:24, Alla Rozovskaya wrote:
            Hello,

            I have built an interpolated count-based LM on the Google
            Web N-gram corpus using SRILM toolkit, as specified here:
            http://www.speech.sri.com/projects/srilm/manpages/srilm-faq.7.html

            Is it possible to use it in moses? In particular, since
            this model uses count files and a file specifying
            weights, what is the right way to specify the path in
            moses.ini?

            Thank you,

            Alla



            _______________________________________________
            Moses-support mailing list
            [email protected]  <mailto:[email protected]>
            http://mailman.mit.edu/mailman/listinfo/moses-support

-- Hieu Hoang
            Researcher
            New York University, Abu Dhabi
            http://www.hoang.co.uk/hieu





--
Hieu Hoang
Researcher
New York University, Abu Dhabi
http://www.hoang.co.uk/hieu

_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to