Github user yanboliang commented on the pull request:

    https://github.com/apache/spark/pull/11247#issuecomment-186040385
  
    @dbtsai Thanks for clarification, I understand the consideration of the 
current implementation. 
    I know that if features are not standardized at all, it may result 
convergency issue when the features have very different scales. This is aligned 
with setting ```standardization = true``` as default value.
    But to some special case(such as the test case which I mentioned above), 
disable standardization will accelerate training. I think the users should have 
the ability to control it.
    Further more, I think enable or disable ```standardization``` will run the 
same route may make users confused. Users may look forward to different result 
or convergency rate(even if it's worse in some cases) for enable or disable 
```standardization```, but they got the same one.
    The only issue we should concern about is that if this change will break 
someone's working training. I think most of the working trainings may trained 
with regularization, it will has little effect. Even if users train without 
regularization, enable or disable ```standardization``` will produce the same 
result currently, users have high possibility to use the default setting.
    I'm still open to hear your thoughts.
    
    BTW, I send #11258 to fix the bug for ```LogisticRegressionWithLBFGS```.
     


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to