Github user Myasuka commented on the pull request:
https://github.com/apache/spark/pull/1290#issuecomment-121966507
hi, @avulanov , I have forked your repository about ann-benchmark
https://github.com/avulanov/ann-benchmark/blob/master/spark/spark.scala . I
feel a little confused about the mini-batch training, it seems that `batchSize`
in code `val trainer = new FeedForwardTrainer(topology, 780,
10).setBatchSize(batchSize)` means the size of sub-block matrix you group the
original input matrix into, and `setMiniBatchFraction(1.0)` in
`trainer.SGDOptimizer.setNumIterations(numIterations).setMiniBatchFraction(1.0).setStepSize(0.03)`
means you actually use full-batch gradient descent not the mini-batch gradient
descent method. Does it performs well on mnist8m data? Maybe you can share the
training parameters in detail, such as layer units, mini-batch size, stepsize
and so on.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]