Hello Jian,

NPLM reports the log-likelihood of the whole training set, and the number is plausible.

assuming you have a minibatch size of 1000, your training set perplexity is exp(1.38122e+08/52853/1000)=13.64

you probably want to measure perplexity on a held-out development set though, with softmax normalisation instead of NCE.

best wishes,
Rico

On 21.07.2015 13:15, jian zhang wrote:
Hi all,

I am running experiments on Bilingual Neural LM.

For extract_training.py, I set
--prune-target-vocab 10000 --prune-source-vocab 10000 --target-context 5 --source-context 4

For train_nplm.py, I set
--ngram-size 14 --output-embedding 512 --input-embedding 192 --hidden 512 --e 5

I use 2 million parallel sentence pairs for training. The implementation is from https://github.com/rsennrich/nplm

From the training log generated by train_nplm.py at the first 2 iterations, I have

Number of training minibatches: 52853
Epoch 1
Current learning rate: 1
Training minibatches: 10000...20000...30000...40000...50000...done.
Training NCE log-likelihood: -1.64277e+08
Writing model
Epoch 2
Current learning rate: 1
Training minibatches: 10000...20000...30000...40000...50000...done.
Training NCE log-likelihood: -1.38122e+08
Writing model


The NCE log-likelihood number is suspicious. It is very low. Did I set any parameters wrong?

Regards,

Jian



--
Jian Zhang
Centre for Next Generation Localisation (CNGL) <http://www.cngl.ie/index.html>
Dublin City University <http://www.dcu.ie/>


_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

_______________________________________________
Moses-support mailing list
[email protected]
http://mailman.mit.edu/mailman/listinfo/moses-support

Reply via email to