On Fri, Nov 04, 2011 at 07:22:15PM +0100, Andreas Müller wrote: > > >> Of course there are many other possibilities like pretraining, > >> deeper networks, different learning rate schedules etc.. > >> You are right, this is somewhat of an active research field > >> Though I have not seen conclusive evidence that any > >> of these methods are consistently better than a vanilla mlp. > > http://www.dumitru.ca/files/publications/icml_07.pdf the table on page 7 > > makes a pretty compelling case, I'd say. > > > These numbers are weired. > A basic grid search with rbf svm gives 1.4% error on mnist.
This was on only 10,000 examples from MNIST (1000 digits per class). Back in 2007, SVM solvers weren't very fast, so they scaled back the problem a bit. David ------------------------------------------------------------------------------ RSA(R) Conference 2012 Save $700 by Nov 18 Register now http://p.sf.net/sfu/rsa-sfdev2dev1 _______________________________________________ Scikit-learn-general mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/scikit-learn-general
