On Tue, Apr 17, 2012 at 03:46:10PM +0200, Andreas Mueller wrote:
> I agree that they show that scaling C seems better.

> BUT: I would not agree with Gael that scale_C=False is broken.

> Even with few samples, it is very hard to actually generate the problem.
> You need to have a learning problem that is VERY sensitive to the value
> of C and you need to have a difference in size of the validation set that
> is larger than the tolerance you have for C.

For the logistic regression with l1 penalty, this happens very easily.
For the logistic regression with l1 penalty, it is not as bad. In
practice, in our problems, it seems that for SVMs, as long as C is 'big
enough', there is not catastrophic failure, though things might be
slightly suboptimal. I realize that the last sentence is in contradiction
with the comon wisdom in statistics that tells you that you are better
off with over-penalizing than under-penalizing.

We found the problem using a logistic-l1 with n_samples ~ 200 and
n_features ~ 50000.

What do people think about my solution 'scale_params'? I thought that it
was a way to make everybody happy, but I don't seem to be getting
traction.

I'd like us to give a good thought to this problem, as I think that it can
be a recurrent pain. I'd actually be happy delaying the release a couple
weeks and reaching a solution that we believe actually solves the problem
for all classes of users (beginners and experts, large and small
n_samples).

Gaƫl

------------------------------------------------------------------------------
Better than sec? Nothing is better than sec when it comes to
monitoring Big Data applications. Try Boundary one-second 
resolution app monitoring today. Free.
http://p.sf.net/sfu/Boundary-dev2dev
_______________________________________________
Scikit-learn-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to