On Wed, Jan 8, 2020 at 9:43 PM <josef.p...@gmail.com> wrote:

>
>
> On Wed, Jan 8, 2020 at 9:38 PM lampahome <pahome.c...@mirlab.org> wrote:
>
>>
>>
>> Stuart Reynolds <stu...@stuartreynolds.net> 於 2020年1月9日 週四 上午10:33寫道:
>>
>>> Correlated features typically have the property that they are tending to
>>> be similarly predictive of the outcome.
>>>
>>> L1 and L2 are both a preference for low coefficients.
>>> If a coefficient can be reduced yet another coefficient maintains
>>> similar loss, the these regularization methods prefer this solution.
>>> If you use L1 or L2, you should mean and variance normalize your
>>> features.
>>>
>>>
>> You mean LASSO and RIDGE both solve multilinearity?
>>
>
> LASSO has the reputation not to be good when there is multicollinearity,
> that's why elastic net L1 + L2 was introduced, AFAIK
>
> With multicollinearity the length of the parameter vector, beta' beta, is
> too large and L2, Ridge shrinks it.
>

e.g.
Marquardt, Donald W., and Ronald D. Snee. "Ridge regression in practice." *The
American Statistician* 29, no. 1 (1975): 3-20.

I just went through it last week because of a argument about variance
inflation factor in Ridge




>
> Josef
>
>
>
>>
>> _______________________________________________
>> scikit-learn mailing list
>> scikit-learn@python.org
>> https://mail.python.org/mailman/listinfo/scikit-learn
>>
>
_______________________________________________
scikit-learn mailing list
scikit-learn@python.org
https://mail.python.org/mailman/listinfo/scikit-learn

Reply via email to