Re: Connecting to Spark cluster remotely

2019-04-22 Thread Andrew Melo
Hi Rishkesh

On Mon, Apr 22, 2019 at 4:26 PM Rishikesh Gawade
 wrote:
>
> To put it simply, what are the configurations that need to be done on the 
> client machine so that it can run driver on itself and executors on 
> spark-yarn cluster nodes?

TBH, if it were me, I would simply SSH to the cluster and start the
spark-shell there. I don't think there's any special spark
configuration you need, but depending on what address space your
cluster is using/where you're connecting from, it might be really hard
to get all the networking components lined up.

>
> On Mon, Apr 22, 2019, 8:22 PM Rishikesh Gawade  
> wrote:
>>
>> Hi.
>> I have been experiencing trouble while trying to connect to a Spark cluster 
>> remotely. This Spark cluster is configured to run using YARN.
>> Can anyone guide me or provide any step-by-step instructions for connecting 
>> remotely via spark-shell?
>> Here's the setup that I am using:
>> The Spark cluster is running with each node as a docker container hosted on 
>> a VM. It is using YARN for scheduling resources for computations.
>> I have a dedicated docker container acting as a spark client, on which i 
>> have the spark-shell installed(spark binary in standalone setup) and also 
>> the Hadoop and Yarn config directories set so that spark-shell can 
>> coordinate with the RM for resources.
>> With all of this set, i tried using the following command:
>>
>> spark-shell --master yarn --deploy-mode client
>>
>> This results in the spark-shell giving me a scala-based console, however, 
>> when I check the Resource Manager UI on the cluster, there seems to be no 
>> application/spark session running.
>> I have been expecting the driver to be running on the client machine and the 
>> executors running in the cluster. But that doesn't seem to happen.
>>
>> How can I achieve this?
>> Is whatever I am trying feasible, and if so, a good practice?
>>
>> Thanks & Regards,
>> Rishikesh

-
To unsubscribe e-mail: user-unsubscr...@spark.apache.org



Re: Connecting to Spark cluster remotely

2019-04-22 Thread Rishikesh Gawade
To put it simply, what are the configurations that need to be done on the
client machine so that it can run driver on itself and executors on
spark-yarn cluster nodes?

On Mon, Apr 22, 2019, 8:22 PM Rishikesh Gawade 
wrote:

> Hi.
> I have been experiencing trouble while trying to connect to a Spark
> cluster remotely. This Spark cluster is configured to run using YARN.
> Can anyone guide me or provide any step-by-step instructions for
> connecting remotely via spark-shell?
> Here's the setup that I am using:
> The Spark cluster is running with each node as a docker container hosted
> on a VM. It is using YARN for scheduling resources for computations.
> I have a dedicated docker container acting as a spark client, on which i
> have the spark-shell installed(spark binary in standalone setup) and also
> the Hadoop and Yarn config directories set so that spark-shell can
> coordinate with the RM for resources.
> With all of this set, i tried using the following command:
>
> spark-shell --master yarn --deploy-mode client
>
> This results in the spark-shell giving me a scala-based console, however,
> when I check the Resource Manager UI on the cluster, there seems to be no
> application/spark session running.
> I have been expecting the driver to be running on the client machine and
> the executors running in the cluster. But that doesn't seem to happen.
>
> How can I achieve this?
> Is whatever I am trying feasible, and if so, a good practice?
>
> Thanks & Regards,
> Rishikesh
>