Hi I think you are mixing things a bit.
Worker is part of the cluster. So it is governed by cluster manager. If you are running standalone cluster, then you can modify spark-env and configure SPARK_WORKER_PORT. executors, on the other hand, are bound with an application, ie spark context. Thus you modify executor properties through a context. So, master != driver and executor != worker. Best Ayan On Fri, May 15, 2015 at 7:52 PM, James King <jakwebin...@gmail.com> wrote: > So I'm using code like this to use specific ports: > > val conf = new SparkConf() > .setMaster(master) > .setAppName("namexxx") > .set("spark.driver.port", "51810") > .set("spark.fileserver.port", "51811") > .set("spark.broadcast.port", "51812") > .set("spark.replClassServer.port", "51813") > .set("spark.blockManager.port", "51814") > .set("spark.executor.port", "51815") > > My question now is : Will the master forward the spark.executor.port value > (to use) to the worker when it hands it a task to do? > > Also the property spark.executor.port is different from the Worker spark > port, how can I make the Worker run on a specific port? > > Regards > > jk > > > On Wed, May 13, 2015 at 7:51 PM, James King <jakwebin...@gmail.com> wrote: > >> Indeed, many thanks. >> >> >> On Wednesday, 13 May 2015, Cody Koeninger <c...@koeninger.org> wrote: >> >>> I believe most ports are configurable at this point, look at >>> >>> http://spark.apache.org/docs/latest/configuration.html >>> >>> search for ".port" >>> >>> On Wed, May 13, 2015 at 9:38 AM, James King <jakwebin...@gmail.com> >>> wrote: >>> >>>> I understated that this port value is randomly selected. >>>> >>>> Is there a way to enforce which spark port a Worker should use? >>>> >>> >>> > -- Best Regards, Ayan Guha