Thank you, Sebastian. This is what I want to understand. Considering the
final score, e.g., accuracy, does this mean that with scaling and without I
will get different results for NB and KNN? Or results will be the same like
in case of decision trees?

With gradient descent algorithms it is clear why I need to scale the
features (because as you wrote for convergence). The question is whether
there are similar reasons to scale features for other algorithms (like I
said, KNN, NB or SVM)? May I get different results (e.g., accuracy) if I
scale features or not?


Best Regards,
Yury Zhauniarovich

On 5 June 2015 at 19:58, Sebastian Raschka <se.rasc...@gmail.com> wrote:

> "Need" to be scaled sounds a little bit strong ;) -- feature scaling is
> really context-dependend. If you are using stochastic gradient descent of
> gradient descent you surely want to standardize your data or at least
> center it for technical reasons and convergence. However, in naive Bayes,
> you just estimate the parameters e.g., via MLE so that there is no
> technical advantage of feature scaling, however, the results will be
> different with and without scaling.
>
> On Jun 5, 2015, at 1:03 PM, Andreas Mueller <t3k...@gmail.com> wrote:
>
>  The result of scaled an non-scaled data will be different because the
> regularization will have a different effect.
>
> On 06/05/2015 03:10 AM, Yury Zhauniarovich wrote:
>
> Thank you all! However, what Sturla wrote is now out of my understanding.
>
>  One more question. It seems also to me that Naive Bayes classifiers also
> do not need data to be scaled. Am I correct?
>
>
> Best Regards,
> Yury Zhauniarovich
>
> On 4 June 2015 at 20:55, Sturla Molden <sturla.mol...@gmail.com> wrote:
>
>> On 04/06/15 20:38, Sturla Molden wrote:
>>
>> > Component-wise EM (aka CEM2) is a better way of avoiding the singularity
>> > disease, though.
>>
>> The traditional EM for a GMM proceeds like this:
>>
>> while True:
>>
>>     global_estep(clusters)
>>
>>     for c in clusters:
>>         mstep(c)
>>
>> This is inherently unstable. Several clusters can become
>> near-singular in the M-step before there is an E-step
>> to redistribute the weights. You can get a "cascade of
>> singularities" where the whole GMM basically dies. Even
>> if you bias the diagonal of the covariance you still
>> have the basic algorithmic problem.
>>
>> CEM2 proceeds like this:
>>
>> while True:
>>     for c in clusters:
>>         estep(c)
>>         mstep(c)
>>
>> This improves stability enormously. When a cluster becomes
>> singular, the memberships are immediately redistributed.
>> Therefore you will not get a "cascade of singularities"
>> where the whole GMM basically dies.
>>
>>
>> Sturla
>>
>>
>>
>> ------------------------------------------------------------------------------
>> _______________________________________________
>> Scikit-learn-general mailing list
>> Scikit-learn-general@lists.sourceforge.net
>> https://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>>
>
>
>
> ------------------------------------------------------------------------------
>
>
>
> _______________________________________________
> Scikit-learn-general mailing 
> listScikit-learn-general@lists.sourceforge.nethttps://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>
>
>
> ------------------------------------------------------------------------------
> _______________________________________________
> Scikit-learn-general mailing list
> Scikit-learn-general@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>
>
>
>
> ------------------------------------------------------------------------------
>
> _______________________________________________
> Scikit-learn-general mailing list
> Scikit-learn-general@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/scikit-learn-general
>
>
------------------------------------------------------------------------------
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to