Re: [Scikit-learn-general] CV scores vs scores on a manual split

2015-02-20 Thread Mathieu Blondel
On Fri, Feb 20, 2015 at 6:57 AM, Andy t3k...@gmail.com wrote: You give the roc_auc_score the result of predict. You should give it the result of predict_proba. This came up already quite a bit, not sure how we can avoid people making this mistake. We can encourage people to use the scorer

[Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread shalu jhanwar
Hi guys, I am using SVM and Random forest classifiers from scikit learn. I wonder is it possible to plot the decision boundary of the model on my own training dataset so that I can have a feeling of the data? Is there any in-built example available in Scikit which I can refer to view let's say

[Scikit-learn-general] ANN: SciPy Latin América 2015 - Call for Proposals

2015-02-20 Thread Raniere Silva
*Call for Proposals* *SciPy Latin América 2015*, the third annual Scientific Computing with Python Conference, will be held this *May 20-22* in *Posadas, Misiones, Argentina*. SciPy is a community dedicated to the advancement of scientific computing through open source Python software for

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread shalu jhanwar
Hi Sebastian, Thanks a lot for your reply. Here in the examples, only 2 features are used to generate these plots. i) Can I do it with more features (I have 16 features)? ii) I wanna see the decision boundary of my training and testing dataset to see if the model is fine or it's overfitted on

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread Sebastian Raschka
Hi, Shalu, One example for plotting decision regions would be here: http://scikit-learn.org/stable/auto_examples/plot_classifier_comparison.html It's basically a brute force approach: You define 2D grid of points and then classifier each of those points. Also, the downside is that you can only

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread Gael Varoquaux
On Fri, Feb 20, 2015 at 05:27:12PM +0100, shalu jhanwar wrote: i) Can I do it with more features (I have 16 features)? How do you visualize a 16-features space? G -- Download BIRT iHub F-Type - The Free

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread shalu jhanwar
Generally I do PCA and can plot the reduced dimension of the data (PC1 and PC2). Here I'm interested in knowing the boundary decision of the classifier. S. On Fri, Feb 20, 2015 at 6:34 PM, Gael Varoquaux gael.varoqu...@normalesup.org wrote: On Fri, Feb 20, 2015 at 05:27:12PM +0100, shalu

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread Sebastian Raschka
i) That would be quite a challenge for the human brain: In the best case you have a hyperplane in 16 dimensions :). How can be put that into a scatter plot!? :) ii + iii) When I understand correctly, you want to get an idea about the generalization error? The simplest way would maybe to look

Re: [Scikit-learn-general] Regarding viewing the decision boundaries of classifiers

2015-02-20 Thread ragv ragv
iii) What would be the best way to know whether the model is fine or overfitted according to your experience? Take a look at this answer by Lars - http://stackoverflow.com/a/12254521/4016687 -- Download BIRT iHub F-Type