[ 
https://issues.apache.org/jira/browse/OPENNLP-155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13023201#comment-13023201
 ] 

Jörn Kottmann commented on OPENNLP-155:
---------------------------------------

I did a little more testing, it turned out that this change increased the 
training time on my test set from 9 minutes to 16 minutes.
I believe its better to have this fix because then the stoping criteria works 
correctly and avoids over training of the model, right?

Anyway I think it should not change the accuracy of the trained model, I did in 
both cases 100 iterations and the difference during evaluation was 2 percent. I 
would like to find out why we get this difference, and not an exactly identical 
model. The stoping criteria didn't terminate the training in my two runs, so I 
think it should not be linked to over training.


> unreliable training set accuracy in perceptron
> ----------------------------------------------
>
>                 Key: OPENNLP-155
>                 URL: https://issues.apache.org/jira/browse/OPENNLP-155
>             Project: OpenNLP
>          Issue Type: Improvement
>          Components: Maxent
>    Affects Versions: maxent-3.0.1-incubating
>            Reporter: Jason Baldridge
>            Assignee: Jason Baldridge
>            Priority: Minor
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>
> The training accuracies reported during perceptron training were much higher 
> than final training accuracy, which turned out to be an artifact of the way 
> training examples were ordered.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to