Hi, Is this the correct link to this open source product?
Armada - how to run millions of batch jobs over thousands of compute nodes using Kubernetes | G-Research <https://www.gresearch.com/news/armada-how-to-run-millions-of-batch-jobs-over-thousands-of-compute-nodes-using-kubernetes/> I am familiar with some of your work in G-Research HTH Dr Mich Talebzadeh, Architect | Data Science | Financial Crime | Forensic Analysis | GDPR view my Linkedin profile <https://www.linkedin.com/in/mich-talebzadeh-ph-d-5205b2/> On Thu, 6 Feb 2025 at 23:40, Dejan Pejchev <de...@gr-oss.io> wrote: > Hello Spark community! > > My name is Dejan Pejchev, and I am a Software Engineer working at > G-Research, and I am a maintainer of our Kubernetes multi-cluster batch > scheduler called Armada. > > We are trying to build an integration with Spark, where we would like to > use the spark-submit with a master armada://xxxx, which will then submit > the driver and executor jobs to Armada. > > I understood the concept of the ExternalClusterManager and how I can write > and provide a new implementation, but I am not clear how can I extend Spark > to accept it. > > I see that in SparkSubmit.scala there is a check for master URLs and it > fails if it isn't any of local, mesos, k8s and yarn. > > What is the correct approach for my use case? > > Thanks in advance, > Dejan Pejchev >