If you are using multiple phrase-tables and generation tables, I don't think there's much you can do about the speed of the decoding. Also, the translation quality is good with this configuration
You can have a look analysis on page 40 here: http://statmt.org/~s0565741/download/ddd.pdf Instead of --translation-factors 0-0+1-1 --generation-factors 1-0 --decoding-steps t0,t1,g0 you're better off doing --translation-factors 0,1-0,1 On 17 April 2013 17:03, Sriram venkatapathy <[email protected]>wrote: > > Thanks Philipp. > > I had tried with very tight t-table limits, even with a limit of 1 for > both words and pos tags and still it didn't work for this example sequence. > This was surprising. > > I hope I can avoid shorter phrase-lengths because the task I have in mind > would require me to have default phrase lengths at least at the POS tag > level. And, would like to avoid using factored model as backoff too because > I would like to encourage those translations that have a particular pattern > of POS tags. > > - Sriram > > > On Tue, Apr 16, 2013 at 9:29 PM, Philipp Koehn <[email protected]>wrote: > >> Hi, >> >> the translation option expansion of factored models may explode in the >> setup that you use above >> (there are many possible lemma and pos mappings, and the product of >> them is explored during >> your first two decoding steps). >> >> You could remedy this by: >> - use shorter phrase lengths >> - use tighter t-table limits >> - use the factored model only as backoff >> >> -phi >> >> On Mon, Apr 15, 2013 at 4:15 PM, Sriram venkatapathy >> <[email protected]> wrote: >> > >> > The decoder (with a factored model) seems to get stuck for certain >> > sentences. For example, >> > >> > It gets stuck for the sentence : >> > >> > saint|noun mary|noun immaculate|adj catholic|nadj church|noun >> > >> > While working without any problem for the following sentences : >> > >> > saint|noun mary|noun immaculate|adj catholic|noun church|noun >> > saint|noun mary|noun immaculate|adj large|nadj church|noun >> > saint|noun mary|noun immaculate|adj large|adj church|noun >> > >> > >> > Here are the training parameters, >> > --translation-factors 0-0+1-1 --generation-factors 1-0 --decoding-steps >> > t0,t1,g0 >> > >> > Factor 0 in both source and target are words >> > Factor 1 in both source and target and part-of-speech tags >> > >> > Any suggestions about what I should be looking at to identify the >> problem ? >> > In the verbose mode, I see that for the problem sentences, the stage of >> > 'collection of translation options' doesn't finish. >> > >> > Thanks ! >> > - Sriram >> > >> > >> > _______________________________________________ >> > Moses-support mailing list >> > [email protected] >> > http://mailman.mit.edu/mailman/listinfo/moses-support >> > >> > > > _______________________________________________ > Moses-support mailing list > [email protected] > http://mailman.mit.edu/mailman/listinfo/moses-support > > -- Hieu Hoang Research Associate University of Edinburgh http://www.hoang.co.uk/hieu
_______________________________________________ Moses-support mailing list [email protected] http://mailman.mit.edu/mailman/listinfo/moses-support
