On Tue, Mar 15, 2016 at 01:01:51PM +0530, Srinivas Kumar wrote: > Hi Ryan, > > My name is Srinivas and I'm a Masters student at the Supercomputer > Education Research Center (India). > > I've worked with different frameworks like OpenMP, MPI, Hadoop and CUDA . I > also have good knowledge in C++, JAVA and Python. > > I'm interested in the "Parallel Stochastic Optimization Methods" project > that you have offered. > > I realized what you are actually looking for is to fine-tune the > implementation of SGD for multi-core rather than extend it to distributed > memory systems. > > I came across a paper by L ́eon Bottou titled "Stochastic Gradient Descent > Tricks" > > Link to paper : http://research.microsoft.com/pubs/192769/tricks-2012.pdf > Link to code : https://github.com/npinto/bottou-sgd > > It offers a nice set of recommendations for implementing SGD. I would like > to know if it would be a good idea to use these recommendations to > implement and fine-tune SGD for *mlpack* ? > > I would like to work on mlpack as a GSoC project. Please let me know what > you think about this idea.
Hi Srinivas, You're right that we are looking to extend the optimizers we have implemented to the multi-core case instead of going fully distributed. You are more than welcome to submit a proposal to implement these tricks (or just implement the tricks and submit a PR for them). When you prepare your proposal, though, you may want to consider more work than just this paper, because I think maybe this paper will not be enough work for a whole summer. Thanks, Ryan -- Ryan Curtin | "Moo." [email protected] | - Eugene Belford _______________________________________________ mlpack mailing list [email protected] https://mailman.cc.gatech.edu/mailman/listinfo/mlpack
