[
https://issues.apache.org/jira/browse/SPARK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Sean Owen resolved SPARK-1454.
------------------------------
Resolution: Won't Fix
> PySpark accumulators fail to update when runJob takes serialized/captured
> closures
> ----------------------------------------------------------------------------------
>
> Key: SPARK-1454
> URL: https://issues.apache.org/jira/browse/SPARK-1454
> Project: Spark
> Issue Type: Bug
> Components: PySpark, Spark Core
> Affects Versions: 1.0.0
> Reporter: William Benton
> Priority: Minor
>
> My patch for SPARK-729 optionally serializes closures when they are cleaned
> (in order to capture the values of mutable free variables at declaration time
> rather than at execution time). This behavior is currently disabled for the
> closure argument to SparkContext.runJob, because enabling it there causes
> Python accumulators to fail to update.
> The purpose of this JIRA is to note this issue and fix whatever is causing
> Python accumulators to behave this way so that closures passed to runJob can
> be captured in general.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]