By default Moses produces translation with ONE factor only. So you can - either modify your reference removing the second factor - or set Moses so that it outputs the second factor too I think the way to do that is through the following parameter -report-all-factors
best, nicola ________________________________________ From: [EMAIL PROTECTED] [EMAIL PROTECTED] On Behalf Of Philipp Koehn [EMAIL PROTECTED] Sent: Friday, June 06, 2008 1:35 PM To: Amit Cc: moses-support Subject: Re: [Moses-support] Mert for factored models Hi, does anybody have an idea what is going on here? -phi On Thu, Jun 5, 2008 at 9:24 PM, Amit <[EMAIL PROTECTED]> wrote: > Hi, > > I am again sorry for bugging you. I looked at the config file and did some > changes and nothing helped. To me the config file looks ok. Its almost > similar to the one on factored model tutorial on SMT website. If it is > possible and doesn't take much time of yours can you please have a look at > the config file and can suggest the part which could go wrong. Its before > mert where the system is dying. > > I am running standard script for tuning. > > perl > /home/srini/MT/Translation/Moses/bin/moses-scripts/scripts-20071216-1351/training/mert-moses.pl > \ > $TUNING/input \ > $TUNING/reference \ > /home/srini/MT/Translation/Moses/bin/moses \ > $MYDIR/model/moses.ini \ > --working-dir $TUNING \ > --rootdir > /home/srini/MT/Translation/Moses/bin/moses-scripts/scripts-20071216-1351 > > Error: > > Start loading LanguageModel > /uusoc/scratch/rome/res/nlp/factored-model/factored-corpus/europarl.lm : > [0.000] seconds > Start loading LanguageModel > /uusoc/scratch/rome/res/nlp/factored-model/factored-corpus/supertag.lm : > [14.000] seconds > Finished loading LanguageModels : [16.000] seconds > Start loading PhraseTable > /uusoc/scratch/rome/res/nlp/factored-model/tuning/filtered/phrase-table.0-0,1.1 > : [16.000] seconds > [ERROR] Malformed input at > Expected input to have words composed of 2 factor(s) (form FAC1|FAC2|...) > but instead received input with 1 factor(s). > > head -2 input > sin embargo , señor presidente , también es realmente necesario que en > biarritz se vaya un poco más lejos . > nosotros , los representantes , tenemos al mismo tiempo el deber de > estimular el progreso , a pesar de la adversidad , y de transmitir los > mensajes que recibimos de la opinión pública en cada uno de nuestros países > . > > head -2 reference > what|WP i|NN would|MD also|RB call|VB for|IN ,|_COMMA_ however|RB ,|_COMMA_ > is|VBZ to|TO look|VB beyond|IN immediate|JJ concerns|NNS in|IN biarritz|NN > .|_PERIOD_ > we|PRP ,|_COMMA_ as|IN elected|VBN representatives|NNS ,|_COMMA_ are|VBP > at|IN least|JJS as|IN responsible|JJ for|IN encouraging|VBG it|PRP to|TO > make|VB progress|NN in|IN the|DT face|NN of|IN adversity|NN as|IN we|PRP > are|VBP for|IN relaying|VBG the|DT messages|NNS that|IN we|PRP receive|VBP > from|IN public|JJ opinion|NN in|IN each|DT of|IN our|PRP$ countries|NNS > .|_PERIOD_ > > Thanks, > Amit > > Philipp Koehn wrote: >> >> Hi, >> >> there may be a few bugs in the current training script, >> so you should check manually that the definition of factors >> in the configuration file matches the data that you use >> as developments set. In your case, you say that the input >> uses only one factor. Check, if the configuration file also >> just specifies one factor. >> >> -phi >> >> On Tue, Jun 3, 2008 at 3:18 PM, Amit <[EMAIL PROTECTED]> wrote: >> >>> >>> Hi Philipp, >>> >>> Sorry for bothering you. I am currently working with Srini at AT&T on >>> SMT. I >>> am trying to run mert for factored models. The way I am doing is the same >>> as >>> I did for unfactored models though now I am getting an error saying >>> input >>> to have words composed of 2 factor. However in my model on source side i >>> have one factor and on target side I have two factors. So the above error >>> doesn't make sense to me. Do I need to run mert differently or am I doing >>> something stupid? >>> >>> Thanks, >>> Amit >>> >>> >>> > > > ######################### > ### MOSES CONFIG FILE ### > ######################### > > # input factors > [input-factors] > 0 > > # mapping steps > [mapping] > 0 T 0 > > # translation tables: source-factors, target-factors, number of scores, file > [ttable-file] > 0 0,1 5 /home/amitg/supertag-model/pos-model/model/phrase-table.0-0,1.gz > > # no generation models, no generation-file section > > # language models: type(srilm/irstlm), factors, order, file > [lmodel-file] > 0 0 5 /home/amitg/supertag-model/lm/english/europarl.lm > 0 1 5 /home/amitg/supertag-model/pos-model/data/pos.lm > > > # limit on how many phrase translations e for each phrase f are loaded > # 0 = all elements loaded > [ttable-limit] > 20 > 0 > # distortion (reordering) weight > [weight-d] > 0.6 > > # language model weights > [weight-l] > 0.2500 > 0.2500 > > > # translation model weights > [weight-t] > 0.2 > 0.2 > 0.2 > 0.2 > 0.2 > > # no generation models, no weight-generation section > > # word penalty > [weight-w] > -1 > > [distortion-limit] > 6 > > _______________________________________________ Moses-support mailing list [email protected] http://mailman.mit.edu/mailman/listinfo/moses-support _______________________________________________ Moses-support mailing list [email protected] http://mailman.mit.edu/mailman/listinfo/moses-support
