On Wed, Jan 8, 2020 at 9:38 PM lampahome <pahome.c...@mirlab.org> wrote:

>
>
> Stuart Reynolds <stu...@stuartreynolds.net> 於 2020年1月9日 週四 上午10:33寫道:
>
>> Correlated features typically have the property that they are tending to
>> be similarly predictive of the outcome.
>>
>> L1 and L2 are both a preference for low coefficients.
>> If a coefficient can be reduced yet another coefficient maintains similar
>> loss, the these regularization methods prefer this solution.
>> If you use L1 or L2, you should mean and variance normalize your features.
>>
>>
> You mean LASSO and RIDGE both solve multilinearity?
>

LASSO has the reputation not to be good when there is multicollinearity,
that's why elastic net L1 + L2 was introduced, AFAIK

With multicollinearity the length of the parameter vector, beta' beta, is
too large and L2, Ridge shrinks it.

Josef



>
> _______________________________________________
> scikit-learn mailing list
> scikit-learn@python.org
> https://mail.python.org/mailman/listinfo/scikit-learn
>
_______________________________________________
scikit-learn mailing list
scikit-learn@python.org
https://mail.python.org/mailman/listinfo/scikit-learn

Reply via email to