Hello Spark community! My name is Dejan Pejchev, and I am a Software Engineer working at G-Research, and I am a maintainer of our Kubernetes multi-cluster batch scheduler called Armada.
We are trying to build an integration with Spark, where we would like to use the spark-submit with a master armada://xxxx, which will then submit the driver and executor jobs to Armada. I understood the concept of the ExternalClusterManager and how I can write and provide a new implementation, but I am not clear how can I extend Spark to accept it. I see that in SparkSubmit.scala there is a check for master URLs and it fails if it isn't any of local, mesos, k8s and yarn. What is the correct approach for my use case? Thanks in advance, Dejan Pejchev