Thanks! 

Regarding the LemmatizerME getting stuck. Should i open an issue?

Markus

-----Original message-----
> From:Rodrigo Agerri <rodrigo.age...@ehu.eus>
> Sent: Wednesday 26th February 2020 13:47
> To: users@opennlp.apache.org
> Subject: Re: LemmatizerTrainerME gets stuck with GaussianSmoothing enabled
> 
> https://opennlp.apache.org/docs/1.8.0/manual/opennlp.html#tools.lemmatizer.training
> 
> The example of training is with a perceptron. An example (with default
> features) can be found here:
> 
> https://github.com/apache/opennlp/blob/master/opennlp-tools/lang/ml/PerceptronTrainerParams.txt
> 
> Best,
> 
> R
> 
> On Wed, 26 Feb 2020 at 11:02, Markus Jelsma <markus.jel...@openindex.io> 
> wrote:
> >
> > Hello Rodrigo,
> >
> > Although i am using the Java API mainly, i use the CLI tools for the 
> > training jobs. And although there is some sporadic mention of Perceptron in 
> > the manual, i have not been able to find any CLI tool that supports 
> > training a Perceptron type model.
> >
> > How do you do it? I use v1.9.2.
> >
> > Thanks,
> > Markus
> >
> > -----Original message-----
> > > From:Rodrigo Agerri <rage...@apache.org>
> > > Sent: Friday 21st February 2020 16:42
> > > To: users@opennlp.apache.org
> > > Subject: Re: LemmatizerTrainerME gets stuck with GaussianSmoothing enabled
> > >
> > > Hello,
> > >
> > > I do not know why is this happening, but while testing the lemmatizer
> > > the best performance (word and sentence accuracy) was always obtained
> > > using the Perceptron trainer (for a number of languages), so I would
> > > recommend to train perceptron models (perhaps you have already tried
> > > this, but just in case).
> > >
> > > Best,
> > >
> > > R
> > >
> > >
> > >
> > >
> > >
> > >
> > > On Fri, 21 Feb 2020 at 16:07, Markus Jelsma <markus.jel...@openindex.io> 
> > > wrote:
> > > >
> > > > Hello,
> > > >
> > > > When GaussianSmoothing is enabled, the LemmatizerTrainerME gets stuck. 
> > > > The loglikelihood gets warped to NaN right after the first iteration, 
> > > > like so:
> > > > Computing model parameters in 16 threads...
> > > > Performing 10000 iterations.
> > > >   1:  ... loglikelihood=-1238383.3965208859     0.8032247485201534
> > > >   2:  ... loglikelihood=NaN     0.8178054720724305
> > > >   3:  ... loglikelihood=NaN     0.8032247485201534
> > > >   4:  ... loglikelihood=NaN     0.8032247485201534
> > > >
> > > > Is this known to happen? Did i stumble upon some bug?
> > > >
> > > > Many thanks,
> > > > Markus
> > >

Reply via email to