Hi Luca.
If you give write comment permissions, I could comment on the google doc in-place which might be helpful. As I think was commented earlier, the current PLS already implements NIPALS. What would the addition be?
Use that in PCA? That is not super clear from the proposal.
I think implementing this together with the other paper you mention will take more than one or two weeks. Please keep in mind that it needs tests, documentation, examples and reviews.

The "massive parallel" paper only has 8 citations since 2013. That seems pretty low impact and not very established. Unsupervised Feature Selection Using Feature Similarity seems a much safer bet (800 cites since 2002), though I am not familiar enough with the area to say if it is still comparable to state of the art or useful. Feature Subset Selection and Ranking for Data Dimensionality Reduction seems borderline with 120 cites since 2007. I haven't actually had time to check the papers (yet?), this is just a first very superficial review.

Instead of focusing on many algorithms, I think you should also allocate some time to ensure that we have good evaluation metrics and cross-validation support for multi-output algorithms where Y might be an input to transform (not sure for how many of these algorithms this is the case).

How is the multi-task lasso that you are proposing different from the one implemented already in scikit-learn?
http://scikit-learn.org/dev/modules/generated/sklearn.linear_model.MultiTaskLasso.html#sklearn.linear_model.MultiTaskLasso

The project sounds great, the hardest part might be finding the right mentor (Gael?)

Cheers,
Andy


On 03/06/2015 07:57 PM, Luca Puggini wrote:
Thanks a lot for the material provided on randomized pca and random forest it would for sure help me in my research.

I talked with my supervisor and he said that I am free to apply for this summer project.

I used quiet a lot GAM and I did some work related to high dimensional fault detection system and so to metrics but apparently these topics are already taken.

My understanding from the previous emails is that nipals PCA may be of interest. On the same topic I have a couple of algorithms that I think could be useful.

1- Sparse principal component analysis via regularized low rank matrix approximation. http://www.sciencedirect.com/science/article/pii/S0047259X07000887 This is basically the equivalent of the nipals algorithm for SPCA. It is more efficient for high dimensional problem. It is pretty useful because it is possible to avoid the initial SVD.

2- Feature Subset Selection and Ranking for Data Dimensionality Reduction http://eprints.whiterose.ac.uk/1947/1/weihl3.pdf .

This is a method to do unsupervised features selection. Similar to SPCA but it is optimized in order to maximize the percentage of explained variance respect to the number of selected variables.


If these topics are not of interest I will be happy to work on
- improve GMM  or -Global optimization based Hyperparameter optimization

I am not familiar with these 2 topics but they are close to my research area so I will be happy to study them.


Now my understanding is that the staff should contact me to discuss further the various arguments. Please fill free to contact me to my private email and I am happy to share my cv and my python code (research quality code )


Thanks a lot,
Luca




------------------------------------------------------------------------------
Dive into the World of Parallel Programming The Go Parallel Website, sponsored
by Intel and developed in partnership with Slashdot Media, is your hub for all
things parallel software development, from weekly thought leadership blogs to
news, videos, case studies, tutorials and more. Take a look and join the
conversation now. http://goparallel.sourceforge.net/


_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

------------------------------------------------------------------------------
Dive into the World of Parallel Programming The Go Parallel Website, sponsored
by Intel and developed in partnership with Slashdot Media, is your hub for all
things parallel software development, from weekly thought leadership blogs to
news, videos, case studies, tutorials and more. Take a look and join the 
conversation now. http://goparallel.sourceforge.net/
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to