Netstat gives exactly the expected IP address (not a 127...., but a 192...). I tried it anyway, though... exactly the same results, but with a number instead of a name. Oh, and I forgot to mention last time, in case it makes a difference - I'm running 0.8.1, not 0.9.0, at least for now
On Sat, Feb 22, 2014 at 12:50 AM, Mayur Rustagi <mayur.rust...@gmail.com>wrote: > most likely the master is binding to a unique address and you are > connecting to some other internal address. Master can bind to random > internal address 127.0... or even your machine IP at that time. > Easiest is to check > netstat -an |grep 7077 > This will give you which IP to bind to exactly when launching spark > context. > > Mayur Rustagi > Ph: +919632149971 > h <https://twitter.com/mayur_rustagi>ttp://www.sigmoidanalytics.com > https://twitter.com/mayur_rustagi > > > > On Fri, Feb 21, 2014 at 9:36 PM, Nathan Kronenfeld < > nkronenf...@oculusinfo.com> wrote: > >> Can anyone help me here? >> >> I've got a small spark cluster running on three machines - hadoop-s1, >> hadoop-s2, and hadoop-s3 - with s1 acting master, and all three acting as >> workers. It works fine - I can connect with spark-shell, I can run jobs, I >> can see the web ui. >> >> The web UI says: >> Spark Master at spark://hadoop-s1.oculus.local:7077 >> URL: spark://hadoop-s1.oculus.local:7077 >> >> I've connected to it fine using both a scala and a java SparkContext. >> >> But when I try connecting from within a Tomcat service, I get the >> following messages: >> [INFO] 22 Feb 2014 00:27:38 - org.apache.spark.Logging$class - Connecting >> to master spark://hadoop-s1.oculus.local:7077... >> [INFO] 22 Feb 2014 00:27:58 - org.apache.spark.Logging$class - Connecting >> to master spark://hadoop-s1.oculus.local:7077... >> [ERROR] 22 Feb 2014 00:28:18 - org.apache.spark.Logging$class - All >> masters are unresponsive! Giving up. >> [ERROR] 22 Feb 2014 00:28:18 - org.apache.spark.Logging$class - Spark >> cluster looks dead, giving up. >> [ERROR] 22 Feb 2014 00:28:18 - org.apache.spark.Logging$class - Exiting >> due to error from cluster scheduler: Spark cluster looks down >> >> When I look on the spark server logs, there isn't even a sign of an >> attempted connection. >> >> I'm trying to use a JavaSparkContext, and I've printed out the parameters >> I pass in, and they work fine in a stand-alone program. >> >> Anyone have a clue why this fails? Or even how to find out why this fals? >> >> >> -- >> Nathan Kronenfeld >> Senior Visualization Developer >> Oculus Info Inc >> 2 Berkeley Street, Suite 600, >> Toronto, Ontario M5A 4J5 >> Phone: +1-416-203-3003 x 238 >> Email: nkronenf...@oculusinfo.com >> > > -- Nathan Kronenfeld Senior Visualization Developer Oculus Info Inc 2 Berkeley Street, Suite 600, Toronto, Ontario M5A 4J5 Phone: +1-416-203-3003 x 238 Email: nkronenf...@oculusinfo.com