On 02/11/2015 04:22 PM, Timothy Vivian-Griffiths wrote: > Hi Gilles, > > Thank you so much for clearing this up for me. So, am I right in thinking > that the feature selection is carried for every CV-fold, and then once the > best parameters have been found, the pipeline is then run on the whole > training set in order to get the .best_estimator_? Yes. > > One final thing, I did manage to find out which of the predictors were being > chosen for the .best_estimator_ but it was not immediately clear how to do > it. In the end, I isolated them by doing the following: > > chosen_predictors = grid2.best_estimator_.steps[0][1].get_support() > I don't think there is an easier way to do this. If you have any idea how to expose this easier, suggestions welcome ;)
------------------------------------------------------------------------------ Dive into the World of Parallel Programming. The Go Parallel Website, sponsored by Intel and developed in partnership with Slashdot Media, is your hub for all things parallel software development, from weekly thought leadership blogs to news, videos, case studies, tutorials and more. Take a look and join the conversation now. http://goparallel.sourceforge.net/ _______________________________________________ Scikit-learn-general mailing list Scikit-learn-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/scikit-learn-general