rdblue commented on a change in pull request #28864:
URL: https://github.com/apache/spark/pull/28864#discussion_r446318223
##
File path:
core/src/main/scala/org/apache/spark/storage/BlockManagerMessages.scala
##
@@ -24,37 +24,37 @@ import org.apache.spark.util.Utils
private[spark] object BlockManagerMessages {
//
- // Messages from the master to slaves.
+ // Messages from the master to storage endpoints.
//
- sealed trait ToBlockManagerSlave
+ sealed trait ToBlockManagerStorage
Review comment:
Nit: I would prefer having `Endpoint` added because I don't think
`Storage` is the correct noun.
##
File path: core/src/main/scala/org/apache/spark/scheduler/DAGScheduler.scala
##
@@ -1912,7 +1912,7 @@ private[spark] class DAGScheduler(
* modify the scheduler's internal state. Use executorLost() to post a loss
event from outside.
*
* We will also assume that we've lost all shuffle blocks associated with
the executor if the
- * executor serves its own blocks (i.e., we're not using external shuffle),
the entire slave
+ * executor serves its own blocks (i.e., we're not using external shuffle),
the entire agent
Review comment:
Should this also be updated to "executor process"?
##
File path:
core/src/main/scala/org/apache/spark/storage/BlockManagerMasterEndpoint.scala
##
@@ -656,7 +655,7 @@ private[spark] class BlockManagerInfo(
var originalLevel: StorageLevel = StorageLevel.NONE
if (blockExists) {
- // The block exists on the slave already.
+ // The block exists on the replica already.
Review comment:
Here's another where the replacement term was changed and you may want
to update the comment.
##
File path:
resource-managers/mesos/src/main/scala/org/apache/spark/scheduler/cluster/mesos/MesosClusterScheduler.scala
##
@@ -731,7 +731,7 @@ private[spark] class MesosClusterScheduler(
override def reregistered(driver: SchedulerDriver, masterInfo: MasterInfo):
Unit = {
logInfo(s"Framework re-registered with master ${masterInfo.getId}")
}
- override def slaveLost(driver: SchedulerDriver, slaveId: SlaveID): Unit = {}
+ override def slaveLost(driver: SchedulerDriver, agentId: SlaveID): Unit = {}
Review comment:
Can we rename the `SlaveID` type to `AgentID` when it is imported?
##
File path:
core/src/test/scala/org/apache/spark/scheduler/DAGSchedulerSuite.scala
##
@@ -466,7 +466,7 @@ class DAGSchedulerSuite extends SparkFunSuite with
LocalSparkContext with TimeLi
assertDataStructuresEmpty()
}
- test("All shuffle files on the slave should be cleaned up when slave lost") {
+ test("All shuffle files on the replica should be cleaned up when it is
lost") {
Review comment:
I think this should be "storage endpoint".
##
File path: core/src/test/scala/org/apache/spark/MapOutputTrackerSuite.scala
##
@@ -136,39 +136,39 @@ class MapOutputTrackerSuite extends SparkFunSuite {
masterTracker.trackerEndpoint =
rpcEnv.setupEndpoint(MapOutputTracker.ENDPOINT_NAME,
new MapOutputTrackerMasterEndpoint(rpcEnv, masterTracker, conf))
-val slaveRpcEnv = createRpcEnv("spark-slave", hostname, 0, new
SecurityManager(conf))
-val slaveTracker = new MapOutputTrackerWorker(conf)
-slaveTracker.trackerEndpoint =
- slaveRpcEnv.setupEndpointRef(rpcEnv.address,
MapOutputTracker.ENDPOINT_NAME)
+val mapWorkerRpcEnv = createRpcEnv("spark-replica", hostname, 0, new
SecurityManager(conf))
Review comment:
What about `spark-worker`?
##
File path:
core/src/main/scala/org/apache/spark/storage/BlockManagerStorageEndpoint.scala
##
@@ -27,17 +27,17 @@ import org.apache.spark.util.{ThreadUtils, Utils}
/**
* An RpcEndpoint to take commands from the master to execute options. For
example,
- * this is used to remove blocks from the slave's BlockManager.
+ * this is used to remove blocks from the replica's BlockManager.
*/
private[storage]
-class BlockManagerSlaveEndpoint(
+class BlockManagerStorageEndpoint(
override val rpcEnv: RpcEnv,
blockManager: BlockManager,
mapOutputTracker: MapOutputTracker)
extends IsolatedRpcEndpoint with Logging {
private val asyncThreadPool =
-
ThreadUtils.newDaemonCachedThreadPool("block-manager-slave-async-thread-pool",
100)
+
ThreadUtils.newDaemonCachedThreadPool("block-manager-replica-async-thread-pool",
100)
Review comment:
Endpoint?
##
File path:
core/src/main/scala/org/apache/spark/storage/BlockManagerStorageEndpoint.scala
##
@@ -27,17 +27,17 @@ import org.apache.spark.util.{ThreadUtils, Utils}
/**
* An RpcEndpoint to take commands from the master to execute options. For
example,
- * this is used to remove blocks from the