Hi devs,

Let me pull some spark-submit developers into this discussion.

@dongjoon-hyun @HyukjinKwon @cloud-fan

What are your thoughts on making spark-submit fully and generically
support ExternalClusterManager implementations?

The current situation is that the only way to submit a Spark job via an
ExternalClusterManager implementation is to use *client mode* and a
*"local" URL*, which seems counterintuitive.

Here is an example for our Armada cluster manager, but this discussion
is meant to be generic for any ExternalClusterManager implementation:

    spark-submit --deploy-mode *client* --master
*local://*armada://IP:PORT ...

What you would expect is this to work:

    spark-submit --deploy-mode cluster --master armada://IP:PORT ...

What are your feelings on making spark-submit support master URLs that
are handled by an ExternalClusterManager implementation in cluster mode?

We are happy to contribute a generic fix for that.

Cheers,
Enrico



Am 07.02.25 um 16:26 schrieb George J:
To me, this seems like a gap in the "pluggable cluster manager"
implementation.

What is the value of making cluster managers pluggable, if
spark-submit doesn't accept jobs on those cluster managers?

It seems to me, for pluggable cluster managers to work, you would want
some parts of spark-submit to be "pluggable" as well.

Thanks!
George Jahad
G-Research Developer

Reply via email to