Re: [Scikit-learn-general] multilayer perceptron questions

2012-06-19 Thread David Marek
Hi On Mon, Jun 18, 2012 at 11:43 AM, Olivier Grisel wrote: > Err, no. The paper I mentioned is even newer: > > http://arxiv.org/abs/1206.1106 > > Just to make it more explicit about the paper content, the title is: > "No More Pesky Learning Rates" and it's a method for estimating the > optimal l

Re: [Scikit-learn-general] multilayer perceptron questions

2012-06-18 Thread Andreas Mueller
Am 18.06.2012 11:43, schrieb Olivier Grisel: > 2012/6/18 Andreas Mueller: >> Hey David. >> Olivier dug up this paper by LeCun's group: >> http://users.ics.aalto.fi/kcho/papers/icml11.pdf >> I think this might be quite interesting for the MLP. > Err, no. The paper I mentioned is even newer: > >h

Re: [Scikit-learn-general] multilayer perceptron questions

2012-06-18 Thread Olivier Grisel
2012/6/18 Andreas Mueller : > Hey David. > Olivier dug up this paper by LeCun's group: > http://users.ics.aalto.fi/kcho/papers/icml11.pdf > I think this might be quite interesting for the MLP. Err, no. The paper I mentioned is even newer: http://arxiv.org/abs/1206.1106 Just to make it more exp

Re: [Scikit-learn-general] multilayer perceptron questions

2012-06-17 Thread Andreas Mueller
Hey David. Olivier dug up this paper by LeCun's group: http://users.ics.aalto.fi/kcho/papers/icml11.pdf I think this might be quite interesting for the MLP. It is probably also interesting for the linear SGD. I'm surprised that they didn't compare against diagonal stochastic Levenberg-Marquardt

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-31 Thread Gael Varoquaux
On Fri, Jun 01, 2012 at 02:49:35AM +0200, David Marek wrote: >I don't have much time these days because I have got exams in school. Good luck! >I have implemented a multi class cross entropy and soft max function and >turned off some of the cython checks, the result is that the cython

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-31 Thread David Marek
Hi, I don't have much time these days because I have got exams in school. I am sorry I haven't informed you. I have implemented a multi class cross entropy and soft max function and turned off some of the cython checks, the result is that the cython implementation is only slightly better, I guess

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-31 Thread Andreas Mueller
Hey David. How is it going? I haven't heard from you in a while. Did you blog anything about your progress? Cheers, Andy Am 16.05.2012 12:15, schrieb David Marek: > On Tue, May 15, 2012 at 4:59 PM, David Warde-Farley > wrote: >> On Tue, May 15, 2012 at 12:12:34AM +0200, David Marek wrote: >>> H

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-17 Thread Justin Bayer
>>>  * Shouldn't the backward step for computing delta_h be: >>>    delta_h[:] = np.dot(delta_o, weights_output.T) * hidden.doutput(x_hidden) >>>    where hidden.doutput is a derivation of the activation function for >>> hidden layer? >> >> Offhand that sounds right. You can use Theano as a sanity

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread David Warde-Farley
On 2012-05-16, at 6:31 AM, Andreas Mueller wrote: > Btw, I am not sure theano is the best way to compute derivatives ;) No? I would agree in the general case. However, in the case of MLPs and backprop, it's a use case for which Theano has been designed and heavily optimized. With it, it's very

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread David Marek
On Tue, May 15, 2012 at 10:31 PM, Andreas Mueller wrote: > On the same topic: I'm not sure if we decided whether we want minibatch, > batch and online learning. > I have the feeling that it might be possible to do particular > optimizations for online learning, and this > is the algorithm that I f

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread Andreas Mueller
Am 16.05.2012 12:29, schrieb David Marek: > Hi > > Yes, I did. I am using gmail so I just quote one mail, didn't want to > answer each mail separately when they are so similar. Sorry, I will > try to be more specific in quoting. > Never mind, probably my mail program just acted up. Btw, I am not su

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread David Marek
Hi Yes, I did. I am using gmail so I just quote one mail, didn't want to answer each mail separately when they are so similar. Sorry, I will try to be more specific in quoting. David 16. 5. 2012 v 12:22, Andreas Mueller : > Hi David. > Did you also see this mail: > http://permalink.gmane.org/gm

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread Andreas Mueller
Hi David. Did you also see this mail: http://permalink.gmane.org/gmane.comp.python.scikit-learn/3071 For some reason it doesn't show up in my inbox and you didn't quote it. So just making sure. Cheers, Andy > Thank you David and Andreas for answering my questions. I will look at Theano.

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread David Marek
On Tue, May 15, 2012 at 4:59 PM, David Warde-Farley wrote: > On Tue, May 15, 2012 at 12:12:34AM +0200, David Marek wrote: >> Hi, >> >> I have worked on multilayer perceptron and I've got a basic >> implementation working. You can see it at >> https://github.com/davidmarek/scikit-learn/tree/gsoc_ml

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread Peter Prettenhofer
2012/5/16 Mathieu Blondel : > > > On Wed, May 16, 2012 at 5:31 AM, Andreas Mueller > wrote: >> >> >> The SequentialDataset was made for vector x vector operations. Depending >> on whether we >> do mini-batch or online learning, the MLP needs vector x matrix or >> matrix x matrix operations. >> In

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-16 Thread Peter Prettenhofer
2012/5/15 Mathieu Blondel : > > > On Tue, May 15, 2012 at 11:59 PM, David Warde-Farley > wrote: >> >> >> I haven't had a look at these classes myself but I think working with raw >> NumPy arrays is a better idea in terms of efficiency. > > > Since it abstracts away the data representation, Sequent

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread Mathieu Blondel
On Wed, May 16, 2012 at 5:31 AM, Andreas Mueller wrote: > > The SequentialDataset was made for vector x vector operations. Depending > on whether we > do mini-batch or online learning, the MLP needs vector x matrix or > matrix x matrix operations. > In particular matrix x matrix is probably not fe

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread Andreas Mueller
On 05/15/2012 10:06 PM, David Warde-Farley wrote: > On 2012-05-15, at 3:23 PM, Andreas Mueller wrote: > >> I am not sure if we want to support sparse data. I have no experience with >> using MLPs on sparse data. >> Could this be done efficiently? The weight vector would need to be >> represented

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread David Warde-Farley
On 2012-05-15, at 3:23 PM, Andreas Mueller wrote: > I am not sure if we want to support sparse data. I have no experience with > using MLPs on sparse data. > Could this be done efficiently? The weight vector would need to be > represented explicitly and densely, I guess. > > Any ideas? People

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread Andreas Mueller
On 05/15/2012 05:16 PM, Mathieu Blondel wrote: On Tue, May 15, 2012 at 11:59 PM, David Warde-Farley mailto:warde...@iro.umontreal.ca>> wrote: I haven't had a look at these classes myself but I think working with raw NumPy arrays is a better idea in terms of efficiency. Since i

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread David Warde-Farley
On Wed, May 16, 2012 at 12:16:21AM +0900, Mathieu Blondel wrote: > On Tue, May 15, 2012 at 11:59 PM, David Warde-Farley < > warde...@iro.umontreal.ca> wrote: > > > > > I haven't had a look at these classes myself but I think working with raw > > NumPy arrays is a better idea in terms of efficiency

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread Mathieu Blondel
On Tue, May 15, 2012 at 11:59 PM, David Warde-Farley < warde...@iro.umontreal.ca> wrote: > > I haven't had a look at these classes myself but I think working with raw > NumPy arrays is a better idea in terms of efficiency. > Since it abstracts away the data representation, SequentialDataset is us

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread David Warde-Farley
On Tue, May 15, 2012 at 12:12:34AM +0200, David Marek wrote: > Hi, > > I have worked on multilayer perceptron and I've got a basic > implementation working. You can see it at > https://github.com/davidmarek/scikit-learn/tree/gsoc_mlp The most > important part is the sgd implementation, which can b

Re: [Scikit-learn-general] multilayer perceptron questions

2012-05-15 Thread Andreas Mueller
Hi David. I'll have a look at your code later today. Let me first answer your questions to my code On 05/15/2012 12:12 AM, David Marek wrote: > Hi, > > 2) I used Andreas' implementation as an inspiration and I am not sure > I understand some parts of it: > * Shouldn't the bias vector be initiali

[Scikit-learn-general] multilayer perceptron questions

2012-05-14 Thread David Marek
Hi, I have worked on multilayer perceptron and I've got a basic implementation working. You can see it at https://github.com/davidmarek/scikit-learn/tree/gsoc_mlp The most important part is the sgd implementation, which can be found here https://github.com/davidmarek/scikit-learn/blob/gsoc_mlp/skl