Hi,

MT Monkey is neural machine translation and not Moses.

Moses does not run on a GPU, it uses only CPU.

When you state that speed is not "real time" what kind of speed are you
looking for?

The best way, as others in this thread have suggested, is to lower the beam
threshold and use the server mode for low latency and multiple cores for
higher throughput.

-phi

On Fri, Dec 23, 2016 at 3:30 PM, Shubham Khandelwal <[email protected]>
wrote:

> Hello,
>
> Currently, I have created one fr-en translation model (size of
> phrase-table.minphr and reordering-table.minlexr are 13 GB and 6.6 GB
> respectively) by following the tutorial of Moses baseline system on a big
> dataset. I have also used Cube Pruning method as suggested by Thomas.
> Now, I use mosesserver and getting response. Now it is taking little bit
> less time to decode the input sentences. However, the decoding is still *not
> *in real time. I have attached moses.ini for your reference.
> To make it fast, I just found an infrastructure: https://
> github.com/ufal/mtmonkey which makes decoding faster by distributed way.
> So, before trying this (mtmonkey) out, I would like to know that Is there
> any other solution or way now by which I can get this decoding in real time
> using Moses ? Is it possible on GPU ?
>
> Looking forward for your response.
>
> Thanking You.
>
> Regards,
> Shubham Khandelwal
>
> On Fri, Dec 16, 2016 at 4:29 PM, Mathias Müller <[email protected]>
> wrote:
>
>> Hi Shubham
>>
>> You could start Moses in server mode:
>>
>> $ moses -f /path/to/moses.ini --server --server-port 12345 --server-log
>> /path/to/log
>>
>> This will load the models, keep them in memory and the server will wait
>> for client requests and serve them until you terminate the process.
>> Translating is a bit different in this case, you have to send an XML-RPC
>> request to the server.
>>
>> But first you'd have to make sure Moses is built with XML-RPC.
>>
>> Regards and good luck
>> Mathias
>> —
>>
>> Mathias Müller
>> AND-2-20
>> Institute of Computational Linguistics
>> University of Zurich
>> Switzerland
>> +41 44 635 75 81 <+41%2044%20635%2075%2081>
>> [email protected]
>>
>> On Fri, Dec 16, 2016 at 10:32 AM, Shubham Khandelwal <[email protected]
>> > wrote:
>>
>>> Hey Thomas,
>>>
>>> Thanks for your reply.
>>> Using Cube Pruning, the speed is littile bit high, but not that much. I
>>> will try to play with these parameters.
>>>
>>> I have binary moses2 which supports it aswell but it is taking more time
>>> to than moses. Can you please send/share somewhere your binary moses2 file
>>> if possible ?
>>>
>>> Also, I do not wish to run this command ( ~/mosesdecoder/bin/moses
>>> -f moses.ini -threads all) every time for every input. Is there any way in
>>> Moses by which all models will load in memory for forever and I can just
>>> pass a input and get output in real time without using this command again
>>> and again.
>>>
>>> Looking forward for your response.
>>>
>>> Thanks again.
>>>
>>> On Fri, Dec 16, 2016 at 1:20 PM, Tomasz Gawryl <
>>> [email protected]> wrote:
>>>
>>>> Hi,
>>>> If you want to speed up decoding time maybe you should consider changing
>>>> searching algorithm. I'm also using compact phrase tables and after some
>>>> test I realised that cube pruning gives almost exactly the same quality
>>>> but
>>>> is much faster. For example you can add something like this to your
>>>> config
>>>> file:
>>>>
>>>> # Cube Pruning
>>>> [search-algorithm]
>>>> 1
>>>> [cube-pruning-pop-limit]
>>>> 1000
>>>> [stack]
>>>> 50
>>>>
>>>>  If your model allows you may also try moses2 binary which is faster
>>>> than
>>>> original.
>>>>
>>>> Regards,
>>>> Thomas
>>>>
>>>> ----------------------------------------------------------------------
>>>>
>>>> Message: 1
>>>> Date: Thu, 15 Dec 2016 19:12:01 +0530
>>>> From: Shubham Khandelwal <[email protected]>
>>>> Subject: Re: [Moses-support] Regarding Decoding Time
>>>> To: Hieu Hoang <[email protected]>
>>>> Cc: moses-support <[email protected]>
>>>> Message-ID:
>>>>         <[email protected]
>>>> ail.com>
>>>> Content-Type: text/plain; charset="utf-8"
>>>>
>>>> Hello,
>>>>
>>>> Currently, I am using phrase-table.minphr , reordering-table.minlexr and
>>>> language model (total size of these 3 are 6 GB). Now, I tried to decode
>>>> on
>>>> two different machines (8 core-16GB RAM  *&* 4 core-40GB RAM) using
>>>> them.
>>>> So, During decoding of around 500 words, it took 90 seconds and 100
>>>> seconds
>>>> respectively on those machines. I am already using compact phrase and
>>>> reordering table representations for faster decoding. Is there any
>>>> other way
>>>> to reduce this decoding time.
>>>>
>>>> Also, In Moses, Do we have distributed way of decoding on multiple
>>>> machines
>>>> ?
>>>>
>>>> Looking forward for your response.
>>>>
>>>> _______________________________________________
>>>> Moses-support mailing list
>>>> [email protected]
>>>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>>>
>>>
>>>
>>>
>>> --
>>> Yours Sincerely,
>>>
>>> Shubham Khandelwal
>>> Masters in Informatics (M2-MoSIG),
>>> University Joseph Fourier-Grenoble INP,
>>> Grenoble, France
>>> Webpage: https://sites.google.com/site/skhandelwl21/
>>>
>>> _______________________________________________
>>> Moses-support mailing list
>>> [email protected]
>>> http://mailman.mit.edu/mailman/listinfo/moses-support
>>>
>>>
>>
>
>
> --
> Yours Sincerely,
>
> Shubham Khandelwal
> Masters in Informatics (M2-MoSIG),
> University Joseph Fourier-Grenoble INP,
> Grenoble, France
> Webpage: https://sites.google.com/site/skhandelwl21/
>
> _______________________________________________
> Moses-support mailing list
> [email protected]
> http://mailman.mit.edu/mailman/listinfo/moses-support
>
>
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to