i think in theory, the parameters don't have to scale, but in practise they have to be otherwise the beam setting will screw up decoding. MERT normalise the L1 weights, eg. these are my weights for a tuned (syntax) model:
weight  L1      L2
0.11    0.11    0.01
0.12    0.12    0.02
-0.03   0.03    0.00
0.03    0.03    0.00
-0.71   0.71    0.50
        1.00    0.53



On 02/05/2010 10:11, Miles Osborne wrote:
there is a large amount of randomness involved with parameter tuning. each time you run it (using the same language resources) you might get different parameters,

also, the parameters are not scaled. this means that one run might give you these values:

10 20 30

and the next run might give you these ones:

0.1 0.2 0.3

Miles

On 2 May 2010 09:34, Somayeh Bakhshaei <[email protected] <mailto:[email protected]>> wrote:


    Hi All,

    A problem:
    Isn't it true that the parameter tuning must gain the structure of
    the language so  i must get the same set of tuned parameters sets
    with different kind of tune sets?
    So why with changing the tuning set i get different amounts for
    parameters?


    another awful result:

    I changed my test set, the Bleu result changed from 19 to 3 !!!!!
    How its may while there is no overlap between none of the test
    sets and train set?!!
    ------------------
    Best Regards,
    S.Bakhshaei



    _______________________________________________
    Moses-support mailing list
    [email protected] <mailto:[email protected]>
    http://mailman.mit.edu/mailman/listinfo/moses-support




--
The University of Edinburgh is a charitable body, registered in Scotland, with registration number SC005336.


_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support
_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to