Yanbo Liang created SPARK-20810:
-----------------------------------
Summary: ML LinearSVC vs MLlib SVMWithSGD output different solution
Key: SPARK-20810
URL: https://issues.apache.org/jira/browse/SPARK-20810
Project: Spark
Issue Type: Question
Components: ML, MLlib
Affects Versions: 2.2.0
Reporter: Yanbo Liang
Fitting with SVM classification model on the same dataset, ML {{LinearSVC}}
produces different solution compared with MLlib {{SVMWithSGD}}. I understand
they use different optimization solver (OWLQN vs SGD), does it make sense to
converge to different solution?
AFAIK, both of them use Hinge loss which is convex but not differentiable
function. Since the derivative of the hinge loss at certain place is
non-deterministic, should we switch to use squared hinge loss?
--
This message was sent by Atlassian JIRA
(v6.3.15#6346)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]