Github user aarondav commented on a diff in the pull request:
https://github.com/apache/spark/pull/2933#discussion_r19377173
--- Diff: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
---
@@ -124,6 +123,10 @@ class DAGScheduler(
/** If enabled, we may run certain actions like take() and first()
locally. */
private val localExecutionEnabled =
sc.getConf.getBoolean("spark.localExecution.enabled", false)
+ /** Broadcast the serialized tasks only when they are bigger than it */
+ private val broadcastTaskMinSize =
+ sc.getConf.getInt("spark.scheduler.broadcastTaskMinSize", 8) * 1024
--- End diff --
Perhaps call this broadcastTaskMinSizeKB? Should we document this flag?
Either way, there should be some mention that your jobs will literally stop
working silently if you change this to be similarly to the akka frame size. It
is not clear that this is sent via Akka.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]