spark-env.sh works as it will be called by spark-submit/spark-shell, or you can just set it before you call spark-shell yourself.
Tim On Tue, Jul 28, 2015 at 1:43 PM, Haripriya Ayyalasomayajula < [email protected]> wrote: > Hi, > > Where can I set the libprocess_ip env variable? spark_env.sh? Thats the > only place I can think of. Can you please point me to any related > documentation? > > On Tue, Jul 28, 2015 at 12:46 PM, Nikolaos Ballas neXus < > [email protected]> wrote: > >> If you are not using any dns like service under /etc/mesos-master/ >> create two files called ip and hostname and put the ip of the eth >> interface. >> >> >> >> Sent from my Samsung device >> >> >> -------- Original message -------- >> From: Haripriya Ayyalasomayajula <[email protected]> >> Date: 28/07/2015 20:18 (GMT+01:00) >> To: [email protected] >> Subject: Problems connecting with Mesos Master >> >> Hi all, >> >> I am trying to use Spark 1.4.1 with Mesos 0.23.0. >> >> When I try to start my spark-shell, it gives me the following warning : >> >> ************************************************** >> >> Scheduler driver bound to loopback interface! Cannot communicate with >> remote master(s). You might want to set 'LIBPROCESS_IP' environment >> variable to use a routable IP address. >> --- >> >> Spark-shell works fine on the node where I run master, but if I start >> running on any of the other slave nodes it gives me the following error: >> >> E0728 11:22:53.176515 10503 socket.hpp:107] Shutdown failed on fd=6: >> Transport endpoint is not connected [107] >> >> E0728 11:22:53.210146 10503 socket.hpp:107] Shutdown failed on fd=6: >> Transport endpoint is not connected [107] >> >> I have the following configs: >> >> >> - zookeeper configured to the mesos master >> - /etc/mesos/zk on all nodes pointing to mesos master ip. >> >> I am not sure if I have to set the ip flag and where I have to set the >> --ip flag? >> >> -- >> Regards, >> Haripriya Ayyalasomayajula >> >> > > > -- > Regards, > Haripriya Ayyalasomayajula > >

