[
https://issues.apache.org/jira/browse/SPARK-2505?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Patrick Wendell updated SPARK-2505:
-----------------------------------
Fix Version/s: (was: 1.1.0)
1.2.0
> Weighted Regularizer
> --------------------
>
> Key: SPARK-2505
> URL: https://issues.apache.org/jira/browse/SPARK-2505
> Project: Spark
> Issue Type: New Feature
> Components: MLlib
> Reporter: DB Tsai
> Fix For: 1.2.0
>
>
> The current implementation of regularization in linear model is using
> `Updater`, and this design has couple issues as the following.
> 1) It will penalize all the weights including intercept. In machine learning
> training process, typically, people don't penalize the intercept.
> 2) The `Updater` has the logic of adaptive step size for gradient decent, and
> we would like to clean it up by separating the logic of regularization out
> from updater to regularizer so in LBFGS optimizer, we don't need the trick
> for getting the loss and gradient of objective function.
> In this work, a weighted regularizer will be implemented, and users can
> exclude the intercept or any weight from regularization by setting that term
> with zero weighted penalty. Since the regularizer will return a tuple of loss
> and gradient, the adaptive step size logic, and soft thresholding for L1 in
> Updater will be moved to SGD optimizer.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]