[
https://issues.apache.org/jira/browse/SPARK-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964524#comment-13964524
]
William Benton commented on SPARK-1454:
---------------------------------------
Can someone please assign this to me?
> PySpark accumulators fail to update when runJob takes serialized/captured
> closures
> ----------------------------------------------------------------------------------
>
> Key: SPARK-1454
> URL: https://issues.apache.org/jira/browse/SPARK-1454
> Project: Spark
> Issue Type: Bug
> Components: PySpark, Spark Core
> Affects Versions: 1.0.0
> Reporter: William Benton
> Priority: Minor
>
> My patch for SPARK-729 optionally serializes closures when they are cleaned
> (in order to capture the values of mutable free variables at declaration time
> rather than at execution time). This behavior is currently disabled for the
> closure argument to SparkContext.runJob, because enabling it there causes
> Python accumulators to fail to update.
> The purpose of this JIRA is to note this issue and fix whatever is causing
> Python accumulators to behave this way so that closures passed to runJob can
> be captured in general.
--
This message was sent by Atlassian JIRA
(v6.2#6252)