Re: [Moses-support] New major release of the continuous space LM toolkit for SMT

2012-06-03 Thread Marcello Federico
I suppose  that an integration is not compatible with the current license of 
CSLM.
GPL cannot  be integrated into LGPL.
Please, correct me if I'm wrong.

Cheers, Marcello

---
Short from my mobile phone

On 04/giu/2012, at 06:12, "Lane Schwartz"  wrote:

> Excellent! Thank you for releasing this, Holger!
> 
> I know you had mentioned that you'd like to get this integrated into
> the decoder. Has anyone from your group been able to work on that?
> 
> Cheers,
> Lane
> 
> 
> On Sun, Jun 3, 2012 at 7:13 PM, Holger Schwenk
>  wrote:
>> I'm happy to announce the availability of a new version of the continuous
>> space
>> language model (CSLM) toolkit.
>> 
>> Continuous space methods we first introduced by Yoshua Bengio in 2001 [1].
>> The basic idea of this approach is to project the word indices onto a
>> continuous space and to use a probability estimator operating on this space.
>> Since the resulting probability functions are smooth functions of the word
>> representation, better generalization to unknown events can be expected.  A
>> neural network can be used to simultaneously learn the projection of the
>> words
>> onto the continuous space and to estimate the n-gram probabilities.  This is
>> still a n-gram approach, but the LM probabilities are interpolated for any
>> possible context of length n-1 instead of backing-off to shorter contexts.
>> 
>> CSLM were initially used in large vocabulary speech recognition systems and
>> more
>> recently in statistical machine translation. Improvements in the perplexity
>> between 10 and 20% relative were reported for many languages and tasks.
>> 
>> 
>> This version of the CSLM toolkit is a major update of the first release. The
>> new features include:
>>  - full support for short-lists during training and inference. By these
>> means,
>>the CSLM can be applied to tasks with large vocabularies.
>>  - very efficient n-best list rescoring.
>>  - support of graphical extension cards (GPU) from Nvidia. This speeds up
>>training by a factor of four with respect to a high-end server with two
>> CPUs.
>> 
>> We successfully trained CSLMs on large tasks like NIST OpenMT'12. Training
>> on one
>> billion words takes less than 24 hours. In our experiments, the CSLM
>> achieves
>> improvements in the BLEU score of up to two points with respect to a large
>> unpruned back-off LM.
>> 
>> A detailed description of the approach can be found in the following
>> publications:
>> 
>> [1] Yoshua Bengio and Rejean Ducharme.  A neural probabilistic language
>> model.
>> In NIPS, vol 13, pages 932--938, 2001.
>> [2] Holger Schwenk, Continuous Space Language Models; in Computer Speech and
>> Language, volume 21, pages 492-518, 2007.
>> [3] Holger Schwenk, Continuous Space Language Models For Statistical Machine
>> Translation; The Prague Bulletin of Mathematical Linguistics, number 83,
>> pages 137-146, 2010.
>> [4] Holger Schwenk, Anthony Rousseau and Mohammed Attik; Large, Pruned or
>> Continuous Space Language Models on a GPU for Statistical Machine
>> Translation,
>> in NAACL workshop on the Future of Language Modeling, June 2012.
>> 
>> 
>> The software is available at http://www-lium.univ-lemans.fr/cslm/. It is
>> distributed under GPL v3.
>> 
>> Comments, bug reports, requests for extensions and contributions are
>> welcome.
>> 
>> enjoy,
>> 
>> Holger Schwenk
>> 
>> LIUM
>> University of Le Mans
>> holger.schw...@lium.univ-lemans.fr
>> 
>> 
>> ___
>> Moses-support mailing list
>> Moses-support@mit.edu
>> http://mailman.mit.edu/mailman/listinfo/moses-support
>> 
> 
> 
> 
> -- 
> When a place gets crowded enough to require ID's, social collapse is not
> far away.  It is time to go elsewhere.  The best thing about space travel
> is that it made it possible to go elsewhere.
> -- R.A. Heinlein, "Time Enough For Love"
> 
> ___
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support

___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] New major release of the continuous space LM toolkit for SMT

2012-06-03 Thread Lane Schwartz
Excellent! Thank you for releasing this, Holger!

I know you had mentioned that you'd like to get this integrated into
the decoder. Has anyone from your group been able to work on that?

Cheers,
Lane


On Sun, Jun 3, 2012 at 7:13 PM, Holger Schwenk
 wrote:
> I'm happy to announce the availability of a new version of the continuous
> space
> language model (CSLM) toolkit.
>
> Continuous space methods we first introduced by Yoshua Bengio in 2001 [1].
> The basic idea of this approach is to project the word indices onto a
> continuous space and to use a probability estimator operating on this space.
> Since the resulting probability functions are smooth functions of the word
> representation, better generalization to unknown events can be expected.  A
> neural network can be used to simultaneously learn the projection of the
> words
> onto the continuous space and to estimate the n-gram probabilities.  This is
> still a n-gram approach, but the LM probabilities are interpolated for any
> possible context of length n-1 instead of backing-off to shorter contexts.
>
> CSLM were initially used in large vocabulary speech recognition systems and
> more
> recently in statistical machine translation. Improvements in the perplexity
> between 10 and 20% relative were reported for many languages and tasks.
>
>
> This version of the CSLM toolkit is a major update of the first release. The
> new features include:
>  - full support for short-lists during training and inference. By these
> means,
>    the CSLM can be applied to tasks with large vocabularies.
>  - very efficient n-best list rescoring.
>  - support of graphical extension cards (GPU) from Nvidia. This speeds up
>    training by a factor of four with respect to a high-end server with two
> CPUs.
>
> We successfully trained CSLMs on large tasks like NIST OpenMT'12. Training
> on one
> billion words takes less than 24 hours. In our experiments, the CSLM
> achieves
> improvements in the BLEU score of up to two points with respect to a large
> unpruned back-off LM.
>
> A detailed description of the approach can be found in the following
> publications:
>
> [1] Yoshua Bengio and Rejean Ducharme.  A neural probabilistic language
> model.
>     In NIPS, vol 13, pages 932--938, 2001.
> [2] Holger Schwenk, Continuous Space Language Models; in Computer Speech and
>     Language, volume 21, pages 492-518, 2007.
> [3] Holger Schwenk, Continuous Space Language Models For Statistical Machine
>     Translation; The Prague Bulletin of Mathematical Linguistics, number 83,
>     pages 137-146, 2010.
> [4] Holger Schwenk, Anthony Rousseau and Mohammed Attik; Large, Pruned or
>     Continuous Space Language Models on a GPU for Statistical Machine
> Translation,
>     in NAACL workshop on the Future of Language Modeling, June 2012.
>
>
> The software is available at http://www-lium.univ-lemans.fr/cslm/. It is
> distributed under GPL v3.
>
> Comments, bug reports, requests for extensions and contributions are
> welcome.
>
> enjoy,
>
> Holger Schwenk
>
> LIUM
> University of Le Mans
> holger.schw...@lium.univ-lemans.fr
>
>
> ___
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
>



-- 
When a place gets crowded enough to require ID's, social collapse is not
far away.  It is time to go elsewhere.  The best thing about space travel
is that it made it possible to go elsewhere.
                -- R.A. Heinlein, "Time Enough For Love"

___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] New major release of the continuous space LM toolkit for SMT

2012-06-03 Thread Holger Schwenk
I'm happy to announce the availability of a new version of the 
continuous space

language model (CSLM) toolkit.

Continuous space methods we first introduced by Yoshua Bengio in 2001 [1].
The basic idea of this approach is to project the word indices onto a
continuous space and to use a probability estimator operating on this space.
Since the resulting probability functions are smooth functions of the word
representation, better generalization to unknown events can be expected.  A
neural network can be used to simultaneously learn the projection of the 
words

onto the continuous space and to estimate the n-gram probabilities.  This is
still a n-gram approach, but the LM probabilities are interpolated for any
possible context of length n-1 instead of backing-off to shorter contexts.

CSLM were initially used in large vocabulary speech recognition systems 
and more

recently in statistical machine translation. Improvements in the perplexity
between 10 and 20% relative were reported for many languages and tasks.


This version of the CSLM toolkit is a major update of the first release. The
new features include:
 - full support for short-lists during training and inference. By these 
means,

   the CSLM can be applied to tasks with large vocabularies.
 - very efficient n-best list rescoring.
 - support of graphical extension cards (GPU) from Nvidia. This speeds up
   training by a factor of four with respect to a high-end server with 
two CPUs.


We successfully trained CSLMs on large tasks like NIST OpenMT'12. 
Training on one
billion words takes less than 24 hours. In our experiments, the CSLM 
achieves

improvements in the BLEU score of up to two points with respect to a large
unpruned back-off LM.

A detailed description of the approach can be found in the following 
publications:


[1] Yoshua Bengio and Rejean Ducharme.  A neural probabilistic language 
model.

In NIPS, vol 13, pages 932--938, 2001.
[2] Holger Schwenk, Continuous Space Language Models; in Computer Speech and
Language, volume 21, pages 492-518, 2007.
[3] Holger Schwenk, Continuous Space Language Models For Statistical Machine
Translation; The Prague Bulletin of Mathematical Linguistics, 
number 83,

pages 137-146, 2010.
[4] Holger Schwenk, Anthony Rousseau and Mohammed Attik; Large, Pruned or
Continuous Space Language Models on a GPU for Statistical Machine 
Translation,

in NAACL workshop on the Future of Language Modeling, June 2012.


The software is available at http://www-lium.univ-lemans.fr/cslm/. It is 
distributed under GPL v3.


Comments, bug reports, requests for extensions and contributions are 
welcome.


enjoy,

Holger Schwenk

LIUM
University of Le Mans
holger.schw...@lium.univ-lemans.fr

___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


Re: [Moses-support] extract translation table

2012-06-03 Thread Barry Haddow
Hi Simon

If you look at the 'extract' file that Moses creates during training,  
it just contains a raw list of all the extracted phrase pairs. If you  
want to find the most common ngram pairs then you could get them from  
this file.

In the phrase table, the phrase counts are normalised to give p(e|f)  
and p(e|f).

cheers - Barry

Quoting Simon Hafner  on Sun, 3 Jun 2012  
21:26:28 +0200:

> Hi all
>
> is there a nice way to get the top 100 translations?
>
> I'm trying to compare two languages on character ngram level, to find
> common edit paths. The idea is to train moses for that pair and then
> extract the most common ngram pairs. Is this even possible or are they
> normalized based on their occurrence?
>
> -- Simon
> ___
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
>
>



-- 
The University of Edinburgh is a charitable body, registered in
Scotland, with registration number SC005336.


___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support


[Moses-support] extract translation table

2012-06-03 Thread Simon Hafner
Hi all

is there a nice way to get the top 100 translations?

I'm trying to compare two languages on character ngram level, to find
common edit paths. The idea is to train moses for that pair and then
extract the most common ngram pairs. Is this even possible or are they
normalized based on their occurrence?

-- Simon
___
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support