[
https://issues.apache.org/jira/browse/OPENNLP-155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13023201#comment-13023201
]
Jörn Kottmann commented on OPENNLP-155:
---------------------------------------
I did a little more testing, it turned out that this change increased the
training time on my test set from 9 minutes to 16 minutes.
I believe its better to have this fix because then the stoping criteria works
correctly and avoids over training of the model, right?
Anyway I think it should not change the accuracy of the trained model, I did in
both cases 100 iterations and the difference during evaluation was 2 percent. I
would like to find out why we get this difference, and not an exactly identical
model. The stoping criteria didn't terminate the training in my two runs, so I
think it should not be linked to over training.
> unreliable training set accuracy in perceptron
> ----------------------------------------------
>
> Key: OPENNLP-155
> URL: https://issues.apache.org/jira/browse/OPENNLP-155
> Project: OpenNLP
> Issue Type: Improvement
> Components: Maxent
> Affects Versions: maxent-3.0.1-incubating
> Reporter: Jason Baldridge
> Assignee: Jason Baldridge
> Priority: Minor
> Original Estimate: 0h
> Remaining Estimate: 0h
>
> The training accuracies reported during perceptron training were much higher
> than final training accuracy, which turned out to be an artifact of the way
> training examples were ordered.
--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira