Github user tgravescs commented on a diff in the pull request:
https://github.com/apache/spark/pull/21729#discussion_r202708012
--- Diff:
core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala ---
@@ -87,7 +87,7 @@ private[spark] class TaskSetManager(
// Set the coresponding index of Boolean var when the task killed by
other attempt tasks,
// this happened while we set the `spark.speculation` to true. The task
killed by others
// should not resubmit while executor lost.
- private val killedByOtherAttempt: Array[Boolean] = new
Array[Boolean](numTasks)
+ private val killedByOtherAttempt = new HashSet[Long]
--- End diff --
@jiangxb1987 please clarify is it fine as is or are you wanting to use a
hashMap and track the index? Can you give an example when this is used for
debugging? For instance are you getting a heap dump and looking at the
datastructures that might make sense, otherwise its not accessible without you
adding in further log statements anyway and its just extra memory usage.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]