Github user thvasilo commented on a diff in the pull request:
https://github.com/apache/spark/pull/4916#discussion_r25937157
--- Diff: ec2/spark_ec2.py ---
@@ -1259,6 +1259,15 @@ def real_main():
cluster_instances=(master_nodes + slave_nodes),
cluster_state='ssh-ready'
)
+
+ # Determine types of running instances
+ existing_master_type = master_nodes[0].instance_type
+ existing_slave_type = slave_nodes[0].instance_type
+ if existing_master_type == existing_slave_type:
+ existing_master_type = ""
+ opts.master_instance_type = existing_master_type
--- End diff --
OK so the use case would be:
1. User starts cluster with some instance type, and later stops it
2. The user now wants to restart the cluster.
3. We can either:
1. Just restart the same instances.
2. Or allow the user to change the instance types by redefining them in
the options passed with the start command.
Is 3.ii a viable use-case? Is it possible to change the type of the
instance being re-launched, for example in the case where the cluster also had
persistent storage .i.e. extra EBS volumes that are still attached to the
stopped instances?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]